Full Text:   <3115>

Summary:  <1761>

CLC number: TP391

On-line Access: 2024-08-27

Received: 2023-10-17

Revision Accepted: 2024-05-08

Crosschecked: 2018-04-12

Cited: 0

Clicked: 7326

Citations:  Bibtex RefMan EndNote GB/T7714

 ORCID:

Zhong-lin Ye

http://orcid.org/0000-0002-2429-3325

-   Go to

Article info.
Open peer comments

Frontiers of Information Technology & Electronic Engineering  2018 Vol.19 No.4 P.524-535

http://doi.org/10.1631/FITEE.1601846


Syntactic word embedding based on dependency syntax and polysemous analysis


Author(s):  Zhong-lin Ye, Hai-xing Zhao

Affiliation(s):  School of Computer Science, Shaanxi Normal University, Xi’an 710119, China; more

Corresponding email(s):   h.x.zhao@163.com

Key Words:  Dependency-based context, Polysemous word representation, Representation learning, Syntactic word embedding



Abstract: 
Most word embedding models have the following problems: (1) In the models based on bag-of-words contexts, the structural relations of sentences are completely neglected; (2) Each word uses a single embedding, which makes the model indiscriminative for polysemous words; (3) Word embedding easily tends to contextual structure similarity of sentences. To solve these problems, we propose an easy-to-use representation algorithm of syntactic word embedding (SWE). The main procedures are: (1) A polysemous tagging algorithm is used for polysemous representation by the latent Dirichlet allocation (LDA) algorithm; (2) Symbols ‘+’ and ‘−’ are adopted to indicate the directions of the dependency syntax; (3) Stopwords and their dependencies are deleted; (4) Dependency skip is applied to connect indirect dependencies; (5) dependency-based contexts are inputted to a word2vec model. Experimental results show that our model generates desirable word embedding in similarity evaluation tasks. Besides, semantic and syntactic features can be captured from dependency-based syntactic contexts, exhibiting less topical and more syntactic similarity. We conclude that SWE outperforms single embedding learning models.

Open peer comments: Debate/Discuss/Question/Opinion

<1>

Please provide your name, email address and a comment





Journal of Zhejiang University-SCIENCE, 38 Zheda Road, Hangzhou 310027, China
Tel: +86-571-87952783; E-mail: cjzhang@zju.edu.cn
Copyright © 2000 - 2025 Journal of Zhejiang University-SCIENCE