Preprocessing techniques

Report a typo

Match the preprocessing techniques of text and their explanations.

Match the items from left and right columns
One-Hot encoding
Count-encoding
TF-IDF vectorization
Word2Vec or FastText vectors
Pretrained vectors from a language model
Trained vectors from scratch
Generates word embeddings based on contextual information
Develops custom embeddings trained on a specific corpus.
Represents token presence in the text using a binary vector with vocabulary size
Utilizes embeddings trained using language modeling techniques
Indicates token frequency in the text using a vector
Creates a vector with weighted token frequencies within the text and corpus
___

Create a free account to access the full topic