Overview
Generative Pre-trained Transformer — a family of autoregressive language models that generate text by predicting the next token.
Cross-References(2)
More in Natural Language Processing
Top-K Sampling
Generation & TranslationA text generation strategy that restricts the model to sampling from the K most probable next tokens.
Text Generation
Generation & TranslationThe process of producing coherent and contextually relevant text using AI language models.
Intent Detection
Generation & TranslationThe classification of user utterances into predefined categories representing the user's goal or purpose, a fundamental component of conversational AI and chatbot systems.
Coreference Resolution
Parsing & StructureThe task of identifying all expressions in text that refer to the same real-world entity.
Speech Recognition
Speech & AudioThe technology that converts spoken language into text, also known as automatic speech recognition.
Information Extraction
Parsing & StructureThe process of automatically extracting structured information from unstructured or semi-structured text sources.
Dialogue Management
Generation & TranslationThe component of conversational systems that tracks conversation state, determines the next system action, and maintains coherent multi-turn interactions with users.
Named Entity Recognition
Parsing & StructureAn NLP task that identifies and classifies named entities in text into categories like person, organisation, and location.