Learning Subword Embedding to Improve Uyghur Named-Entity Recognition
AbstractUyghur is a morphologically rich and typical agglutinating language, and morphological segmentation affects the performance of Uyghur named-entity recognition (NER). Common Uyghur NER systems use the word sequence as input and rely heavily on feature engineering. However, semantic information cannot be fully learned and will easily suffer from data sparsity arising from morphological processes when only the word sequence is considered. To solve this problem, we provide a neural network architecture employing subword embedding with character embedding based on a bidirectional long short-term memory network with a conditional random field layer. Our experiments show that subword embedding can effectively enhance the performance of the Uyghur NER, and the proposed method outperforms the model-based word sequence method. View Full-Text
Share & Cite This Article
Saimaiti, A.; Wang, L.; Yibulayin, T. Learning Subword Embedding to Improve Uyghur Named-Entity Recognition. Information 2019, 10, 139.
Saimaiti A, Wang L, Yibulayin T. Learning Subword Embedding to Improve Uyghur Named-Entity Recognition. Information. 2019; 10(4):139.Chicago/Turabian Style
Saimaiti, Alimu; Wang, Lulu; Yibulayin, Tuergen. 2019. "Learning Subword Embedding to Improve Uyghur Named-Entity Recognition." Information 10, no. 4: 139.
Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.