Overview
The balance between a model's ability to minimise bias (error from assumptions) and variance (sensitivity to training data fluctuations).
More in Machine Learning
SHAP Values
MLOps & ProductionA game-theoretic approach to explaining individual model predictions by computing each feature's marginal contribution, based on Shapley values from cooperative game theory.
Model Registry
MLOps & ProductionA versioned catalogue of trained machine learning models with metadata, lineage, and approval workflows, enabling reproducible deployment and governance at enterprise scale.
Hierarchical Clustering
Unsupervised LearningA clustering method that builds a tree-like hierarchy of clusters through successive merging or splitting of groups.
Continual Learning
MLOps & ProductionA machine learning paradigm where models learn from a continuous stream of data, accumulating knowledge over time without forgetting previously learned information.
Catastrophic Forgetting
Anomaly & Pattern DetectionThe tendency of neural networks to completely lose previously learned knowledge when trained on new tasks, a fundamental challenge in continual and multi-task learning.
Content-Based Filtering
Unsupervised LearningA recommendation approach that suggests items similar to those a user has previously liked, based on item attributes.
Class Imbalance
Feature Engineering & SelectionA situation where the distribution of classes in a dataset is significantly skewed, with some classes vastly outnumbering others.
Ensemble Methods
MLOps & ProductionMachine learning techniques that combine multiple models to produce better predictive performance than any single model, including bagging, boosting, and stacking approaches.