Published January 1, 2018
| Version v1
Conference paper
Open
Characters or Morphemes: How to Represent Words?
Creators
- 1. Middle East Tech Univ, Cognit Sci Dept, Informat Inst, Ankara, Turkey
- 2. Hacettepe Univ, Dept Comp Engn, Ankara, Turkey
Description
In this paper, we investigate the effects of using subword information in representation learning. We argue that using syntactic subword units effects the quality of the word representations positively. We introduce a morpheme-based model and compare it against to word-based, characterbased, and character n-gram level models. Our model takes a list of candidate segmentations of a word and learns the representation of the word based on different segmentations that are weighted by an attention mechanism. We performed experiments on Turkish as a morphologically rich language and English with a comparably poorer morphology. The results show that morpheme-based models are better at learning word representations of morphologically complex languages compared to character-based and character ngram level models since the morphemes help to incorporate more syntactic knowledge in learning, that makes morphemebased models better at syntactic tasks.
Files
bib-db77027f-3048-428a-89e9-05250815ed34.txt
Files
(122 Bytes)
| Name | Size | Download all |
|---|---|---|
|
md5:20b413609bb9bd24a552bf616eae9a3c
|
122 Bytes | Preview Download |