Overview
An unsupervised technique for discovering abstract topics that occur in a collection of documents.
More in Natural Language Processing
Coreference Resolution
Parsing & StructureThe task of identifying all expressions in text that refer to the same real-world entity.
Multilingual Model
Semantics & RepresentationA language model trained on text from dozens or hundreds of languages simultaneously, enabling cross-lingual understanding and generation without language-specific fine-tuning.
Instruction Following
Semantics & RepresentationThe capability of language models to accurately interpret and execute natural language instructions, a core skill developed through instruction tuning and alignment training.
Large Language Model
Semantics & RepresentationA neural network trained on massive text corpora that can generate, understand, and reason about natural language.
Document Understanding
Core NLPAI systems that extract structured information from unstructured documents by combining optical character recognition, layout analysis, and natural language comprehension.
BERT
Semantics & RepresentationBidirectional Encoder Representations from Transformers — a language model that understands context by reading text in both directions.
Text Embedding Model
Core NLPA neural network trained to convert text passages into fixed-dimensional vectors that capture semantic meaning, enabling similarity search, clustering, and retrieval applications.
Context Window
Semantics & RepresentationThe maximum amount of text a language model can consider at once when generating a response.