SENTENCE_SEGMENTER_MODEL¶
-
lexnlp.nlp.en.segments.sentences.
SENTENCE_SEGMENTER_MODEL
= <nltk.tokenize.punkt.PunktSentenceTokenizer object>¶ A sentence tokenizer which uses an unsupervised algorithm to build a model for abbreviation words, collocations, and words that start sentences; and then uses that model to find sentence boundaries. This approach has been shown to work well for many European languages.