Overview
An attention mechanism that selectively computes relationships between a subset of input tokens rather than all pairs, reducing quadratic complexity in transformer models.
Cross-References(2)
More in Artificial Intelligence
AI Safety
Safety & GovernanceThe interdisciplinary field dedicated to making AI systems safe, robust, and beneficial while minimizing risks of unintended consequences.
Strong AI
Foundations & TheoryA theoretical form of AI that would have consciousness, self-awareness, and the ability to truly understand rather than simulate understanding.
Artificial Intelligence
Foundations & TheoryThe simulation of human intelligence processes by computer systems, including learning, reasoning, and self-correction.
Direct Preference Optimisation
Training & InferenceA simplified alternative to RLHF that directly optimises language model policies using preference data without requiring a separate reward model.
AI Interpretability
Safety & GovernanceThe degree to which humans can understand the internal mechanics and reasoning of an AI model's predictions and decisions.
Commonsense Reasoning
Foundations & TheoryThe AI capability to make inferences based on everyday knowledge that humans typically take for granted.
Constraint Satisfaction
Reasoning & PlanningA computational approach where problems are defined as a set of variables, domains, and constraints that must all be simultaneously satisfied.
Semantic Web
Foundations & TheoryAn extension of the World Wide Web that enables machines to interpret and process web content through standardised semantic metadata.
See Also
Transformer
A neural network architecture based entirely on attention mechanisms, eliminating recurrence and enabling parallel processing of sequences.
Deep LearningAttention Mechanism
A neural network component that learns to focus on relevant parts of the input when producing each element of the output.
Deep Learning