Source: r-cran-tokenizers Standards-Version: 4.7.3 Maintainer: Debian R Packages Maintainers Uploaders: Andreas Tille , Section: gnu-r Testsuite: autopkgtest-pkg-r Build-Depends: debhelper-compat (= 13), dh-r, r-base-dev, r-cran-stringi, r-cran-rcpp, r-cran-snowballc, architecture-is-64-bit, architecture-is-little-endian, Vcs-Browser: https://salsa.debian.org/r-pkg-team/r-cran-tokenizers Vcs-Git: https://salsa.debian.org/r-pkg-team/r-cran-tokenizers.git Homepage: https://cran.r-project.org/package=tokenizers Rules-Requires-Root: no Package: r-cran-tokenizers Architecture: any Depends: ${R:Depends}, ${shlibs:Depends}, ${misc:Depends}, Recommends: ${R:Recommends}, Suggests: ${R:Suggests}, Description: GNU R fast, consistent tokenization of natural language text Convert natural language text into tokens. Includes tokenizers for shingled n-grams, skip n-grams, words, word stems, sentences, paragraphs, characters, shingled characters, lines, tweets, Penn Treebank, regular expressions, as well as functions for counting characters, words, and sentences, and a function for splitting longer texts into separate documents, each with the same number of words. The tokenizers have a consistent interface, and the package is built on the 'stringi' and 'Rcpp' packages for fast yet correct tokenization in 'UTF-8'.