Overview
Continuous observation of deployed machine learning models to detect performance degradation, data drift, anomalous predictions, and infrastructure issues in production.
Cross-References(2)
More in Machine Learning
Bias-Variance Tradeoff
Training TechniquesThe balance between a model's ability to minimise bias (error from assumptions) and variance (sensitivity to training data fluctuations).
Transfer Learning
Advanced MethodsA technique where knowledge gained from training on one task is applied to a different but related task.
Bagging
Advanced MethodsBootstrap Aggregating — an ensemble method that trains multiple models on random subsets of data and averages their predictions.
Underfitting
Training TechniquesWhen a model is too simple to capture the underlying patterns in the data, resulting in poor performance on both training and test data.
Gradient Boosting
Supervised LearningAn ensemble technique that builds models sequentially, with each new model correcting residual errors of the combined ensemble.
XGBoost
Supervised LearningAn optimised distributed gradient boosting library designed for speed and performance in machine learning competitions and production.
Support Vector Machine
Supervised LearningA supervised learning algorithm that finds the optimal hyperplane to separate different classes in high-dimensional space.
Elastic Net
Training TechniquesA regularisation technique combining L1 and L2 penalties, balancing feature selection and coefficient shrinkage.