function tokenizing rtext objects
# S3 method for rtext
text_tokenize(
string,
regex = NULL,
ignore.case = FALSE,
fixed = FALSE,
perl = FALSE,
useBytes = FALSE,
non_token = FALSE
)
text to be tokenized
regex expressing where to cut see (see grep)
whether or not reges should be case sensitive (see grep)
whether or not regex should be interpreted as is or as regular expression (see grep)
whether or not Perl compatible regex should be used (see grep)
byte-by-byte matching of regex or character-by-character (see grep)
should information for non-token, i.e. those patterns by which the text was splitted, be returned as well