Overview
Techniques and architectures that enable language models to process and reason over extremely long input sequences, from tens of thousands to millions of tokens.
More in Natural Language Processing
Seq2Seq Model
Core NLPA neural network architecture that maps an input sequence to an output sequence, used in translation and summarisation.
Relation Extraction
Parsing & StructureIdentifying semantic relationships between entities mentioned in text.
Constitutional AI
Core NLPAn approach to AI alignment where models are trained to follow a set of principles or constitution.
Text Embedding
Core NLPDense vector representations of text passages that capture semantic meaning for similarity comparison and retrieval.
Vector Database
Core NLPA database optimised for storing and querying high-dimensional vector embeddings for similarity search.
Structured Output
Semantics & RepresentationThe generation of machine-readable formatted responses such as JSON, XML, or code from language models, enabling reliable integration with downstream software systems.
Reranking
Core NLPA two-stage retrieval process where an initial set of candidate documents is rescored by a more powerful model to improve the relevance ordering of search results.
Topic Modelling
Text AnalysisAn unsupervised technique for discovering abstract topics that occur in a collection of documents.