Overview
A training paradigm where AI models are refined using human preference signals, aligning model outputs with human values and quality expectations through reward modelling.
More in Artificial Intelligence
BLEU Score
Evaluation & MetricsA metric for evaluating the quality of machine-generated text by comparing it to reference translations or texts.
ROC Curve
Evaluation & MetricsA graphical plot illustrating the diagnostic ability of a binary classifier as its discrimination threshold is varied.
System Prompt
Prompting & InteractionAn initial instruction set provided to a language model that defines its persona, constraints, output format, and behavioural guidelines for a given session or application.
AI Guardrails
Safety & GovernanceSafety mechanisms and constraints implemented around AI systems to prevent harmful, biased, or policy-violating outputs while preserving useful functionality.
Model Pruning
Models & ArchitectureThe process of removing redundant or less important parameters from a neural network to reduce its size and computational cost.
Retrieval-Augmented Generation
Infrastructure & OperationsA technique combining information retrieval with text generation, allowing AI to access external knowledge before generating responses.
Few-Shot Prompting
Prompting & InteractionA technique where a language model is given a small number of examples within the prompt to guide its response pattern.
Chain-of-Thought Prompting
Prompting & InteractionA prompting technique that encourages language models to break down reasoning into intermediate steps before providing an answer.