NLP basics
Mar 26, 2025 路 3 min read 路 Cleaning: Pre-processing Tokenization: It is a process to convert paragraphs or sentences into tokens Corpus: ParagraphsDocument: Sentences in the paragraphVocabulary: All the unique words in the documentWords: tokens in the sentences Stemming Reduc...
Join discussion