Skip to content

NLP

Posted on:September 23, 2022 at 03:22 PM

Lexicon Normalization

Bag-of-Words Model :

Convert text to word count vectors with CountVectorizer. Both tokenize a collection of text documents and build a vocabulary of known words

Convert text to word frequency vectors with TfidfVectorizer.

Convert text to unique integers with HashingVectorizer.