Ruby port of the NLTK Punkt sentence segmentation algorithm
-
Updated
Jun 10, 2018 - Ruby
Ruby port of the NLTK Punkt sentence segmentation algorithm
TUFS Asian Language Parallel Corpus
A PHP Library to extract n-grams from a text. Simple preprocessing tools (cleaning, tokenizing) included.
In the same time, flames (such as rants, taunts, and squalid phrases) are offensive/abusive phrases which might attack or offend the users for a variety of reasons. An automatic discriminative software with a sensitivity parameter for flame or abusive language detection would be a useful tool.
Deep Learning Nanodegree Project : To generate Simpsons TV scripts using part of the Simpsons data-set of scripts from 27 seasons.
Add a description, image, and links to the tokenized-sentences topic page so that developers can more easily learn about it.
To associate your repository with the tokenized-sentences topic, visit your repo's landing page and select "manage topics."