NLP basics
Mar 26, 2025 · 3 min read · Cleaning: Pre-processing Tokenization: It is a process to convert paragraphs or sentences into tokens Corpus: ParagraphsDocument: Sentences in the paragraphVocabulary: All the unique words in the documentWords: tokens in the sentences Stemming Reduc...
Join discussion