Overview
A training strategy that presents examples to a model in a meaningful order, typically from easy to hard.
Cross-References(1)
More in Machine Learning
Underfitting
Training TechniquesWhen a model is too simple to capture the underlying patterns in the data, resulting in poor performance on both training and test data.
Label Noise
Feature Engineering & SelectionErrors or inconsistencies in the annotations of training data that can degrade model performance and lead to unreliable predictions if not properly addressed.
Bias-Variance Tradeoff
Training TechniquesThe balance between a model's ability to minimise bias (error from assumptions) and variance (sensitivity to training data fluctuations).
Feature Store
MLOps & ProductionA centralised repository for storing, managing, and serving machine learning features, ensuring consistency between training and inference environments across an organisation.
Boosting
Supervised LearningAn ensemble technique that sequentially trains models, each focusing on correcting the errors of previous models.
Continual Learning
MLOps & ProductionA machine learning paradigm where models learn from a continuous stream of data, accumulating knowledge over time without forgetting previously learned information.
Matrix Factorisation
Unsupervised LearningA technique that decomposes a matrix into constituent matrices, widely used in recommendation systems and dimensionality reduction.
Tabular Deep Learning
Supervised LearningThe application of deep neural networks to structured tabular datasets, competing with traditional methods like gradient boosting through specialised architectures and regularisation.