Overview
The process of converting a trained model into a format that can be stored, transferred, and later reconstructed for inference.
More in Machine Learning
Gradient Descent
Training TechniquesAn optimisation algorithm that iteratively adjusts parameters in the direction of steepest descent of the loss function.
Learning Rate
Training TechniquesA hyperparameter that controls how much model parameters are adjusted with respect to the loss gradient during training.
Matrix Factorisation
Unsupervised LearningA technique that decomposes a matrix into constituent matrices, widely used in recommendation systems and dimensionality reduction.
Backpropagation
Training TechniquesThe algorithm for computing gradients of the loss function with respect to network weights, enabling neural network training.
Anomaly Detection
Anomaly & Pattern DetectionIdentifying data points, events, or observations that deviate significantly from the expected pattern in a dataset.
SMOTE
Feature Engineering & SelectionSynthetic Minority Over-sampling Technique — a method for addressing class imbalance by generating synthetic examples of the minority class.
Loss Function
Training TechniquesA mathematical function that measures the difference between predicted outputs and actual target values during model training.
Model Registry
MLOps & ProductionA versioned catalogue of trained machine learning models with metadata, lineage, and approval workflows, enabling reproducible deployment and governance at enterprise scale.