| |
| |
| import nltk |
| from nltk import pos_tag |
| from nltk.tokenize import word_tokenize |
| from nltk.corpus import stopwords |
| from collections import Counter |
|
|
| |
| nltk.download('punkt') |
| nltk.download('averaged_perceptron_tagger') |
| nltk.download('stopwords') |
|
|
|
|
|
|
| |
| stop_words = set(stopwords.words('english')) |
|
|
| def preprocess(text): |
| tokens = word_tokenize(text.lower()) |
| return [word for word in tokens if word.isalnum() and word not in stop_words] |
|
|
| def get_keywords(text, top_n=5): |
| processed_text = preprocess(text) |
| pos_tags = pos_tag(processed_text) |
| |
| |
| keywords = [word for word, pos in pos_tags if pos.startswith(('NN', 'VB', 'JJ'))] |
| |
| |
| keyword_counts = Counter(keywords) |
| return [word for word, _ in keyword_counts.most_common(top_n)] |