Stemming or lemmatization normalizes words for counting and analysis.
Frequency analysis counts how often a word appears in a text.
N-grams extend frequency analysis to multi-term phrases.
Vectorization represents words/documents as vectors in N-dimensional space to capture relationships.
Extracting key phrases from text helps identify the main terms in NLP.
Data mining workloads focus on searching and indexing large amounts of data.
Knowledge mining is an AI workload that makes large amounts of data searchable.
Conversational AI is part of NLP and facilitates the creation of chatbots.
Language models predict the next word in a sequence of words based on context.