Overview
Bidirectional Encoder Representations from Transformers — a language model that understands context by reading text in both directions.
Cross-References(1)
More in Natural Language Processing
Reranking
Core NLPA two-stage retrieval process where an initial set of candidate documents is rescored by a more powerful model to improve the relevance ordering of search results.
Text Classification
Text AnalysisThe task of assigning predefined categories or labels to text documents based on their content.
Token Limit
Semantics & RepresentationThe maximum number of tokens a language model can process in a single input-output interaction.
Text Summarisation
Text AnalysisThe process of creating a concise and coherent summary of a longer text document while preserving key information.
Context Window
Semantics & RepresentationThe maximum amount of text a language model can consider at once when generating a response.
Top-K Sampling
Generation & TranslationA text generation strategy that restricts the model to sampling from the K most probable next tokens.
Multilingual Model
Semantics & RepresentationA language model trained on text from dozens or hundreds of languages simultaneously, enabling cross-lingual understanding and generation without language-specific fine-tuning.
Extractive Summarisation
Generation & TranslationA summarisation technique that identifies and selects the most important sentences from a source document to compose a condensed version without generating new text.