Overview
A summarisation technique that identifies and selects the most important sentences from a source document to compose a condensed version without generating new text.
More in Natural Language Processing
Seq2Seq Model
Core NLPA neural network architecture that maps an input sequence to an output sequence, used in translation and summarisation.
Large Language Model
Semantics & RepresentationA neural network trained on massive text corpora that can generate, understand, and reason about natural language.
Context Window
Semantics & RepresentationThe maximum amount of text a language model can consider at once when generating a response.
Language Model
Semantics & RepresentationA probabilistic model that assigns probabilities to sequences of words, enabling prediction of the next word in a sequence.
Part-of-Speech Tagging
Parsing & StructureThe process of assigning grammatical categories (noun, verb, adjective) to each word in a text.
Token Limit
Semantics & RepresentationThe maximum number of tokens a language model can process in a single input-output interaction.
Sentiment Analysis
Text AnalysisThe computational study of people's opinions, emotions, and attitudes expressed in text.
GloVe
Semantics & RepresentationGlobal Vectors for Word Representation — an unsupervised learning algorithm for obtaining word vector representations from aggregated word co-occurrence statistics.