The `Tokenzier.transform` method in Python's `pyspark.ml.feature` module is used to tokenize a text column by splitting it into individual words or tokens. This method takes the input DataFrame as a parameter and applies the tokenization process to the specified text column. The output DataFrame will contain an additional column with the tokenized words for each record in the input DataFrame.
Python Tokenizer.transform - 60 examples found. These are the top rated real world Python examples of pyspark.ml.feature.Tokenizer.transform extracted from open source projects. You can rate examples to help us improve the quality of examples.