Exemplo n.º 1
0
class WordTokenizer(Tokenizer):

	def __init__(self):
		"""
		Creates a new WordTokenizer instance
		"""
		self.__preprocessor = Preprocessor()

	def tokenize(self, string):
		"""
		Tokenizes the specified string

		Returns a list of token
		"""
		cleanedText = self.__preprocessor.clean(string)
		
		return cleanedText.split()
Exemplo n.º 2
0
	def __init__(self):
		"""
		Creates a new WordTokenizer instance
		"""
		self.__preprocessor = Preprocessor()