The search functionality is under construction.

IEICE TRANSACTIONS on Information

Contextualized Character Embedding with Multi-Sequence LSTM for Automatic Word Segmentation

Hyunyoung LEE, Seungshik KANG

  • Full Text Views

    0

  • Cite this

Summary :

Contextual information is a crucial factor in natural language processing tasks such as sequence labeling. Previous studies on contextualized embedding and word embedding have explored the context of word-level tokens in order to obtain useful features of languages. However, unlike it is the case in English, the fundamental task in East Asian languages is related to character-level tokens. In this paper, we propose a contextualized character embedding method using n-gram multi-sequences information with long short-term memory (LSTM). It is hypothesized that contextualized embeddings on multi-sequences in the task help each other deal with long-term contextual information such as the notion of spans and boundaries of segmentation. The analysis shows that the contextualized embedding of bigram character sequences encodes well the notion of spans and boundaries for word segmentation rather than that of unigram character sequences. We find out that the combination of contextualized embeddings from both unigram and bigram character sequences at output layer rather than the input layer of LSTMs improves the performance of word segmentation. The comparison showed that our proposed method outperforms the previous models.

Publication
IEICE TRANSACTIONS on Information Vol.E103-D No.11 pp.2371-2378
Publication Date
2020/11/01
Publicized
2020/08/19
Online ISSN
1745-1361
DOI
10.1587/transinf.2020EDP7038
Type of Manuscript
PAPER
Category
Natural Language Processing

Authors

Hyunyoung LEE
  Kookmin University
Seungshik KANG
  Kookmin University

Keyword