Word Level Tokenizer Interface¶
- class konoha.word_tokenizer.WordTokenizer(tokenizer: str = 'MeCab', user_dictionary_path: Optional[str] = None, system_dictionary_path: Optional[str] = None, model_path: Optional[str] = None, mode: Optional[str] = None, dictionary_format: Optional[str] = None, endpoint: Optional[str] = None, ssl: Optional[bool] = None, port: Optional[int] = None)¶