I am trying to tokenize a sentence using nltk. when i do it through python shell i get the correct answer.
>>> import nltk
>>> sentence = "Mohanlal made his acting debut in Thiranottam (1978), but the film got released only after 25 years due to censorship issues."
>>> tokens = nltk.word_tokenize(sentence)
>>> tokens
['Mohanlal', 'made', 'his', 'acting', 'debut', 'in', 'Thiranottam', '(', '1978', ')', ',', 'but', 'the', 'film', 'got', 'released', 'only', 'after', '25', 'years', 'due', 'to', 'censorship', 'issues', '.']
But when i write the same code in a file and try to run it i got the following error.
Traceback (most recent call last):
File "tokenize.py", line 1, in <module>
import nltk
File "/usr/local/lib/python2.7/dist-packages/nltk/__init__.py", line 114, in <module>
from nltk.collocations import *
File "/usr/local/lib/python2.7/dist-packages/nltk/collocations.py", line 38, in <module>
from nltk.util import ngrams
File "/usr/local/lib/python2.7/dist-packages/nltk/util.py", line 13, in <module>
import pydoc
File "/usr/lib/python2.7/pydoc.py", line 55, in <module>
import sys, imp, os, re, types, inspect, __builtin__, pkgutil, warnings
File "/usr/lib/python2.7/inspect.py", line 39, in <module>
import tokenize
File "/home/gadheyan/Project/Codes/tokenize.py", line 2, in <module>
from nltk import word_tokenize
ImportError: cannot import name word_tokenize
Here's the code that i run.
import nltk
from nltk import word_tokenize
sentence = "Mohanlal made his acting debut in Thiranottam (1978), but the film got released only after 25 years due to censorship issues."
tokens = nltk.word_tokenize(sentence)
print tokens