Overview
The maximum number of tokens a language model can process in a single input-output interaction.
Cross-References(1)
More in Natural Language Processing
Vector Database
Core NLPA database optimised for storing and querying high-dimensional vector embeddings for similarity search.
Sentiment Analysis
Text AnalysisThe computational study of people's opinions, emotions, and attitudes expressed in text.
Multilingual Model
Semantics & RepresentationA language model trained on text from dozens or hundreds of languages simultaneously, enabling cross-lingual understanding and generation without language-specific fine-tuning.
Byte-Pair Encoding
Parsing & StructureA subword tokenisation algorithm that iteratively merges the most frequent character pairs to build a vocabulary.
Slot Filling
Core NLPThe task of extracting specific parameter values from user utterances to fulfil a detected intent, such as identifying dates, locations, and names in booking requests.
Machine Translation
Generation & TranslationThe use of AI to automatically translate text or speech from one natural language to another.
Seq2Seq Model
Core NLPA neural network architecture that maps an input sequence to an output sequence, used in translation and summarisation.
Structured Output
Semantics & RepresentationThe generation of machine-readable formatted responses such as JSON, XML, or code from language models, enabling reliable integration with downstream software systems.