I am trying to use Naive Bayes algorithm to do sentimental analysis and was going through a few articles. As mentioned in almost every article I need to train my Naive Bayes algorithm with some pre-computed sentiments.
Now, I have a piece of code using movie_review module provided with NLTK. The code is :
import nltk
import random
from nltk.corpus import movie_reviews
documents = [(list(movie_reviews.words(fileid)), category)
for category in movie_reviews.categories()
for fileid in movie_reviews.fileids(category)]
random.shuffle(documents)
all_words = []
for w in movie_reviews.words():
all_words.append(w.lower())
all_words = nltk.FreqDist(all_words)
word_features = list(all_words.keys())[:3000]
def find_features(document):
words = set(document)
features = {}
for w in word_features:
features[w] = (w in words)
return features
featuresets = [(find_features(rev), category) for (rev, category) in documents]
training_set = featuresets[:1900]
testing_set = featuresets[1900:]
classifier = nltk.NaiveBayesClassifier.train(training_set)
print("Classifier accuracy percent:",(nltk.classify.accuracy(classifier, testing_set))*100)
So, in the above code I have a training_set and a testing_set. I checked the movie_review module and inside the movie review module we have many small text files containing reviews.
- So, my question is here we had the movie review module and we imported it and trained and tested using the module but how can we do when I am using an external training data set and external testing data set.
- Also, how is NLTK parsing movie_review directory which has so many text files inside it. As I will be using http://ai.stanford.edu/~amaas/data/sentiment/ this as my training data set, so I need to understand how its done.