Overview
A text generation strategy that restricts the model to sampling from the K most probable next tokens.
Cross-References(2)
More in Natural Language Processing
Semantic Similarity
Semantics & RepresentationA measure of how closely the meanings of two text passages align, computed through embedding comparison and used in duplicate detection, search, and recommendation systems.
Prompt Injection
Semantics & RepresentationA security vulnerability where malicious inputs manipulate a language model into ignoring its instructions or producing unintended outputs.
Long-Context Modelling
Semantics & RepresentationTechniques and architectures that enable language models to process and reason over extremely long input sequences, from tens of thousands to millions of tokens.
GPT
Semantics & RepresentationGenerative Pre-trained Transformer — a family of autoregressive language models that generate text by predicting the next token.
Slot Filling
Core NLPThe task of extracting specific parameter values from user utterances to fulfil a detected intent, such as identifying dates, locations, and names in booking requests.
Tokenisation
Semantics & RepresentationThe process of breaking text into smaller units (tokens) such as words, subwords, or characters for processing by language models.
Natural Language Understanding
Core NLPThe subfield of NLP focused on machine reading comprehension and extracting meaning from text.
Token Limit
Semantics & RepresentationThe maximum number of tokens a language model can process in a single input-output interaction.