Overview
A simplified alternative to RLHF that directly optimises language model policies using preference data without requiring a separate reward model.
Cross-References(2)
More in Artificial Intelligence
Cognitive Computing
Foundations & TheoryComputing systems that simulate human thought processes using self-learning algorithms, data mining, pattern recognition, and natural language processing.
Retrieval-Augmented Generation
Infrastructure & OperationsA technique combining information retrieval with text generation, allowing AI to access external knowledge before generating responses.
AI Safety
Safety & GovernanceThe interdisciplinary field dedicated to making AI systems safe, robust, and beneficial while minimizing risks of unintended consequences.
Abductive Reasoning
Reasoning & PlanningA form of logical inference that seeks the simplest and most likely explanation for a set of observations.
Symbolic AI
Foundations & TheoryAn approach to AI that uses human-readable symbols and rules to represent problems and derive solutions through logical reasoning.
Artificial Intelligence
Foundations & TheoryThe simulation of human intelligence processes by computer systems, including learning, reasoning, and self-correction.
Bayesian Reasoning
Reasoning & PlanningA statistical approach to AI that uses Bayes' theorem to update probability estimates as new evidence becomes available.
Constraint Satisfaction
Reasoning & PlanningA computational approach where problems are defined as a set of variables, domains, and constraints that must all be simultaneously satisfied.
See Also
Language Model
A probabilistic model that assigns probabilities to sequences of words, enabling prediction of the next word in a sequence.
Natural Language ProcessingRLHF
Reinforcement Learning from Human Feedback — a technique for aligning language models with human preferences through reward modelling.
Natural Language Processing