# -*- coding: utf-8 -*-
"""
nercut 0.2
Dictionary-based maximal matching word segmentation, constrained with
Thai Character Cluster (TCC) boundaries, and combining tokens that are
parts of the same named-entity.
Code by Wannaphong Phatthiyaphaibun
"""
from typing import Iterable, List
from pythainlp.tag.named_entity import NER
_thainer = NER(engine="thainer")
[docs]def segment(
text: str,
taglist: Iterable[str] = [
"ORGANIZATION",
"PERSON",
"PHONE",
"EMAIL",
"DATE",
"TIME",
],
) -> List[str]:
"""
Dictionary-based maximal matching word segmentation, constrained with
Thai Character Cluster (TCC) boundaries, and combining tokens that are
parts of the same named-entity.
:param str text: text to be tokenized to words
:parm list taglist: a list of named-entity tags to be used
:return: list of words, tokenized from the text
"""
if not text or not isinstance(text, str):
return []
global _thainer
tagged_words = _thainer.tag(text, pos=False)
words = []
combining_word = ""
combining_word = ""
for curr_word, curr_tag in tagged_words:
if curr_tag != "O":
tag = curr_tag[2:]
else:
tag = "O"
if curr_tag.startswith("B-") and tag in taglist:
combining_word = curr_word
elif (
curr_tag.startswith("I-")
and combining_word != ""
and tag in taglist
):
combining_word += curr_word
elif (
curr_tag == "O"
and combining_word != ""
):
words.append(combining_word)
combining_word = ""
words.append(curr_word)
else:
combining_word = ""
words.append(curr_word)
return words