step_clean_levels       Clean categorical levels
step_clean_names        Clean variable names
step_lda                Calculates lda dimension estimates
step_lemma              Lemmatization of tokenlist variables
step_ngram              Generate ngrams from tokenlist
step_pos_filter         Part of speech filtering of tokenlist variables
step_sequence_onehot    Generate the basic set of text features
step_stem               Stemming of tokenlist variables
step_stopwords          Filtering of stopwords from a tokenlist
                        variable
step_text_normalization
                        Normalization of tokenlist variables
step_textfeature        Generate the basic set of text features
step_texthash           Term frequency of tokens
step_tf                 Term frequency of tokens
step_tfidf              Term frequency-inverse document frequency of
                        tokens
step_tokenfilter        Filter the tokens based on term frequency
step_tokenize           Tokenization of character variables
step_tokenmerge         Generate the basic set of text features
step_untokenize         Untokenization of tokenlist variables
step_word_embeddings    Pretrained word embeddings of tokens
tokenlist               Create tokenlist object
