Overview
The application of deep neural networks to structured tabular datasets, competing with traditional methods like gradient boosting through specialised architectures and regularisation.
Cross-References(3)
More in Machine Learning
Elastic Net
Training TechniquesA regularisation technique combining L1 and L2 penalties, balancing feature selection and coefficient shrinkage.
Adam Optimiser
Training TechniquesAn adaptive learning rate optimisation algorithm combining momentum and RMSProp for efficient deep learning training.
Model Monitoring
MLOps & ProductionContinuous observation of deployed machine learning models to detect performance degradation, data drift, anomalous predictions, and infrastructure issues in production.
Learning Rate
Training TechniquesA hyperparameter that controls how much model parameters are adjusted with respect to the loss gradient during training.
Model Calibration
MLOps & ProductionThe process of adjusting a model's predicted probabilities so they accurately reflect the true likelihood of outcomes, essential for risk-sensitive decision-making.
Automated Machine Learning
MLOps & ProductionThe end-to-end automation of the machine learning pipeline including feature engineering, model selection, hyperparameter tuning, and deployment, making ML accessible to non-experts.
Feature Selection
MLOps & ProductionThe process of identifying and selecting the most relevant input variables for a machine learning model.
t-SNE
Unsupervised Learningt-Distributed Stochastic Neighbour Embedding — a technique for visualising high-dimensional data in two or three dimensions.