⚠ Deprecated: Legacy
This package has been marked deprecated in favour of the official maven Stanford CoreNLP package. Read more here https://github.com/sergey-tihon/Stanford.NLP.NET.
Tokenization of raw text is a standard pre-processing step for many NLP tasks. For English, tokenization usually involves punctuation splitting and separation of some affixes like possessives. Other languages require more extensive token pre-processing, which is usually called segmentation.
$ dotnet add package Stanford.NLP.SegmenterNo README available.