Graph-based Relation Mining for Context-free Out-of-vocabulary Word Embedding Learning
Ziran Liang, Yuyin Lu, HeGang Chen, Yanghui Rao
Main: Machine Learning for NLP Main-oral Paper
Session 6: Machine Learning for NLP (Oral)
Conference Room: Metropolitan Centre
Conference Time: July 12, 09:00-10:30 (EDT) (America/Toronto)
Global Time: July 12, Session 6 (13:00-14:30 UTC)
Keywords:
graph-based methods, word embeddings
TLDR:
The out-of-vocabulary (OOV) words are difficult to represent while critical to the performance of embedding-based downstream models. Prior OOV word embedding learning methods failed to model complex word formation well. In this paper, we propose a novel graph-based relation mining method, namely GRM...
You can open the
#paper-P3542
channel in a separate window.
Abstract:
The out-of-vocabulary (OOV) words are difficult to represent while critical to the performance of embedding-based downstream models. Prior OOV word embedding learning methods failed to model complex word formation well. In this paper, we propose a novel graph-based relation mining method, namely GRM, for OOV word embedding learning. We first build a Word Relationship Graph (WRG) based on word formation and associate OOV words with their semantically relevant words, which can mine the relational information inside word structures. Subsequently, our GRM can infer high-quality embeddings for OOV words through passing and aggregating semantic attributes and relational information in the WRG, regardless of contextual richness. Extensive experiments demonstrate that our model significantly outperforms state-of-the-art baselines on both intrinsic and downstream tasks when faced with OOV words.