On 03/08/18 08:49, ma...@mail.com wrote:
I like to analyse text. my method consisted of something like
words=text.split(), which would split the text into space-seperated
units. then I tried to use the Python NLTK library, which had alot
of features I wanted, but using `word-tokenize' gives a different
  answer.-

What gives?.

I don't know or use NLTK and you haven't provided any examples, so I can't say for certain. However vague memories from the natural language work I did decades ago suggest that perhaps splitting text by spaces is a little too simplistic, and NLTK may be trying to do something more comprehensive.

--
Rhodri James *-* Kynesim Ltd
--
https://mail.python.org/mailman/listinfo/python-list

Reply via email to