SENTENCE_SEGMENTER_MODEL

lexnlp.nlp.en.segments.sentences.SENTENCE_SEGMENTER_MODEL = <nltk.tokenize.punkt.PunktSentenceTokenizer object>

A sentence tokenizer which uses an unsupervised algorithm to build a model for abbreviation words, collocations, and words that start sentences; and then uses that model to find sentence boundaries. This approach has been shown to work well for many European languages.