Overview
The method of dividing long documents into smaller segments for embedding and retrieval, balancing context preservation with optimal chunk sizes for vector search accuracy.
Cross-References(1)
More in Natural Language Processing
Contextual Embedding
Semantics & RepresentationWord representations that change based on surrounding context, capturing polysemy and contextual meaning.
Code Generation
Semantics & RepresentationThe automated production of source code from natural language specifications or partial code context, powered by large language models trained on programming repositories.
Multilingual Model
Semantics & RepresentationA language model trained on text from dozens or hundreds of languages simultaneously, enabling cross-lingual understanding and generation without language-specific fine-tuning.
Speech Synthesis
Speech & AudioThe artificial production of human speech from text, also known as text-to-speech.
BERT
Semantics & RepresentationBidirectional Encoder Representations from Transformers — a language model that understands context by reading text in both directions.
Structured Output
Semantics & RepresentationThe generation of machine-readable formatted responses such as JSON, XML, or code from language models, enabling reliable integration with downstream software systems.
Token Limit
Semantics & RepresentationThe maximum number of tokens a language model can process in a single input-output interaction.
GPT
Semantics & RepresentationGenerative Pre-trained Transformer — a family of autoregressive language models that generate text by predicting the next token.