GPT
Generative Pre-trained Transformer — a family of large language models developed by OpenAI. GPT models are trained to predict the next token in a sequence and can generate coherent, contextually relevant text across many tasks.
Why It Matters
The GPT series (GPT-3, GPT-4, etc.) popularized LLMs and sparked the current AI revolution. Understanding GPT helps contextualize the broader LLM landscape.
Example
GPT-4 powering ChatGPT to write essays, debug code, analyze images, and hold nuanced conversations across virtually any topic.
Think of it like...
Like a very well-read author who has studied the patterns of millions of texts and can now write convincingly in almost any style or on any subject.
Related Terms
Large Language Model
A type of AI model trained on massive amounts of text data that can understand and generate human-like text. LLMs use transformer architecture and typically have billions of parameters, enabling them to perform a wide range of language tasks.
Transformer
A neural network architecture introduced in 2017 that uses self-attention mechanisms to process sequential data in parallel rather than sequentially. Transformers are the foundation of modern LLMs like GPT, Claude, and Gemini.
OpenAI
The AI research company that created GPT, ChatGPT, DALL-E, and Whisper. Originally founded as a nonprofit in 2015, OpenAI became the most prominent AI company after launching ChatGPT.
ChatGPT
OpenAI's consumer-facing AI chatbot powered by GPT models. ChatGPT brought LLMs to the mainstream when it launched in November 2022, reaching 100 million users in two months.
Pre-training
The initial phase of training a model on a large, general-purpose dataset before specializing it for specific tasks. Pre-training gives the model broad knowledge and capabilities.