File manager - Edit - /usr/local/lib/python3.9/dist-packages/pythainlp/tokenize/pyicu.py
Back
# -*- coding: utf-8 -*- # SPDX-FileCopyrightText: 2016-2024 PyThaiNLP Project # SPDX-License-Identifier: Apache-2.0 """ Wrapper for PyICU word segmentation. This wrapper module uses :class:`icu.BreakIterator` with Thai as :class:`icu.Local` to locate boundaries between words in the text. :See Also: * `GitHub repository <https://github.com/ovalhub/pyicu>`_ """ import re from typing import List from icu import BreakIterator, Locale def _gen_words(text: str) -> str: bd = BreakIterator.createWordInstance(Locale("th")) bd.setText(text) p = bd.first() for q in bd: yield text[p:q] p = q def segment(text: str) -> List[str]: """ :param str text: text to be tokenized into words :return: list of words, tokenized from the text """ if not text or not isinstance(text, str): return [] text = re.sub("([^\u0E00-\u0E7F\n ]+)", " \\1 ", text) return list(_gen_words(text))
| ver. 1.4 |
Github
|
.
| PHP 7.4.33 | Generation time: 0.66 |
proxy
|
phpinfo
|
Settings