Overview
An ensemble technique that builds models sequentially, with each new model correcting residual errors of the combined ensemble.
More in Machine Learning
Gradient Descent
Training TechniquesAn optimisation algorithm that iteratively adjusts parameters in the direction of steepest descent of the loss function.
Loss Function
Training TechniquesA mathematical function that measures the difference between predicted outputs and actual target values during model training.
Backpropagation
Training TechniquesThe algorithm for computing gradients of the loss function with respect to network weights, enabling neural network training.
Ensemble Methods
MLOps & ProductionMachine learning techniques that combine multiple models to produce better predictive performance than any single model, including bagging, boosting, and stacking approaches.
Content-Based Filtering
Unsupervised LearningA recommendation approach that suggests items similar to those a user has previously liked, based on item attributes.
Dimensionality Reduction
Unsupervised LearningTechniques that reduce the number of input variables in a dataset while preserving essential information and structure.
Clustering
Unsupervised LearningUnsupervised learning technique that groups similar data points together based on inherent patterns without predefined labels.
Adam Optimiser
Training TechniquesAn adaptive learning rate optimisation algorithm combining momentum and RMSProp for efficient deep learning training.