nlp - Best tokenization method for dealing with informal english text data? -


in natural langauge processing, there tokenization tools designed accurately tokenization informal english text data sentences? eg: informal sources such reddit comments or forum data.

i've tried stanford tokenizer not seem informal text sources ones mentioned.

with influx of informal text data social media, hoping there more accurate way of tokenizing such data further processing.


Comments

Popular posts from this blog

ubuntu - PHP script to find files of certain extensions in a directory, returns populated array when run in browser, but empty array when run from terminal -

php - How can i create a user dashboard -

javascript - How to detect toggling of the fullscreen-toolbar in jQuery Mobile? -