Overview
Rectified Linear Unit — an activation function that outputs the input directly if positive, otherwise outputs zero.
Cross-References(1)
More in Deep Learning
Parameter-Efficient Fine-Tuning
Language ModelsMethods for adapting large pretrained models to new tasks by only updating a small fraction of their parameters.
Data Parallelism
ArchitecturesA distributed training strategy that replicates the model across multiple devices and divides training data into batches processed simultaneously, synchronising gradients after each step.
Batch Normalisation
ArchitecturesA technique that normalises layer inputs during training to stabilise and accelerate deep neural network learning.
Deep Learning
ArchitecturesA subset of machine learning using neural networks with multiple layers to learn hierarchical representations of data.
Convolutional Layer
ArchitecturesA neural network layer that applies learnable filters across input data to detect local patterns and features.
Encoder-Decoder Architecture
ArchitecturesA neural network design where an encoder processes input into a fixed representation and a decoder generates output from it.
Skip Connection
ArchitecturesA neural network shortcut that allows the output of one layer to bypass intermediate layers and be added to a later layer's output.
Pretraining
ArchitecturesTraining a model on a large general dataset before fine-tuning it on a specific downstream task.