Overview
AI systems designed to engage in natural, context-aware dialogue with humans across multiple turns.
More in Natural Language Processing
Long-Context Modelling
Semantics & RepresentationTechniques and architectures that enable language models to process and reason over extremely long input sequences, from tens of thousands to millions of tokens.
BERT
Semantics & RepresentationBidirectional Encoder Representations from Transformers — a language model that understands context by reading text in both directions.
Structured Output
Semantics & RepresentationThe generation of machine-readable formatted responses such as JSON, XML, or code from language models, enabling reliable integration with downstream software systems.
Byte-Pair Encoding
Parsing & StructureA subword tokenisation algorithm that iteratively merges the most frequent character pairs to build a vocabulary.
Natural Language Generation
Core NLPThe subfield of NLP concerned with producing natural language text from structured data or representations.
GPT
Semantics & RepresentationGenerative Pre-trained Transformer — a family of autoregressive language models that generate text by predicting the next token.
Instruction Tuning
Semantics & RepresentationTraining a language model to follow natural language instructions by fine-tuning on instruction-response pairs.
Relation Extraction
Parsing & StructureIdentifying semantic relationships between entities mentioned in text.