中国综合性科技类核心期刊(北大核心)

中国科学引文数据库来源期刊(CSCD)

美国《化学文摘》(CA)收录

美国《数学评论》(MR)收录

俄罗斯《文摘杂志》收录

Message Board

Respected readers, authors and reviewers, you can add comments to this page on any questions about the contribution, review, editing and publication of this journal. We will give you an answer as soon as possible. Thank you for your support!

Name
E-mail
Phone
Title
Content
Verification Code
Issue 3
May  2021
Turn off MathJax
Article Contents
LIANG Yanchun, FANG Ailian. Chinese text relation extraction based on a multi-channel convolutional neural network[J]. Journal of East China Normal University (Natural Sciences), 2021, (3): 96-104. doi: 10.3969/j.issn.1000-5641.2021.03.010
Citation: LIANG Yanchun, FANG Ailian. Chinese text relation extraction based on a multi-channel convolutional neural network[J]. Journal of East China Normal University (Natural Sciences), 2021, (3): 96-104. doi: 10.3969/j.issn.1000-5641.2021.03.010

Chinese text relation extraction based on a multi-channel convolutional neural network

doi: 10.3969/j.issn.1000-5641.2021.03.010
  • Received Date: 2020-05-18
  • Publish Date: 2021-05-01
  • This paper presents an end-to-end method for Chinese text relation extraction based on a multi-channel CNN (convolutional neural network). Each channel is stacked with a layered neural network; these channels do not interact during recurrent propagation, which enables a neural network to learn different representations. Considering the nuances of the Chinese language, we employed the attention mechanism to extract the semantic features of a sentence, and then integrate structural information using piecewise average pooling. After the maximum pooling layer, the final representation of the sentence is obtained and a relational score is calculated. Finally, the ranking-loss function is used to replace the cross-entropy function for training. The experimental results show that the MCNN_Att_RL (Multi CNN_Att_RL) model proposed in this paper can effectively improve the precision, recall, and F1 value of entity relation extraction.
  • loading
  • [1]
    LIU C Y, SUN W B, CHAO W H, et al. Convolution neural network for relation extraction [C]// International Conference on Advanced Data Mining and Applications. Berlin: Springer, 2013: 231-242.
    [2]
    ZENG D J, LIU K, LAI S W, et al. Relation classification via convolutional deep neural network [C]// International Conference on Computational Linguistocs. 2014: 2335-2344.
    [3]
    ZHANG D, WANG D. Relation classification via recurrent neural network [J]. Computer Ence, 2015, 36(36): 257-266.
    [4]
    ZHOU P, SHI W, TIAN J, et al. Attention-based bidirectional long short-term memory networks for relation classification [C]// Meeting of the Association for Computational Linguistics. 2016: 207-212.
    [5]
    ZHU J Z, QIAO J Z, DAI X X, et al. Relation classification via target-concentrated attention CNNs [C]// International Conference On Neural Information Processing. 2017: 137-146.
    [6]
    LI J, HUANG G M, CHEN J H, et al. Dual CNN for relation extraction with knowledge-based attention and word embeddings [J]. Computational Intelligence and Neuroscience, 2019, 171: 1-10. doi:  10.1155/2019/6789520
    [7]
    HONG Y, LIU Y X, YANG S Z, et al. Improving graph convolutional networks based on relation-aware attention for end-to-end relation extraction [J]. IEEE Access, 2020(8): 51315-51323. doi:  10.1109/ACCESS.2020.2980859
    [8]
    ZENG D J, LIU K, CHEN Y B, et al. Distant supervision for relation extraction via piecewise convolutional neural networks [C]// Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing. 2015: 1753-1762.
    [9]
    LIN Y K, SHEN S Q, LIU Z Y, et al. Neural relation extraction with selective attention over instances [C]// Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2016: 2124-2133.
    [10]
    KIM E K, CHOI K S. Improving distantly supervised relation extraction by knowledge base-driven zero subject resolution [J]. ICE Transactions on Information and Systems, 2018, 26(3): 142-151. doi:  10.1587/transinf.2017EDL8270
    [11]
    SUN T T, ZHANG C H, JI Y, et al. MSnet: Multi-head self-attention network for distantly supervised relation extraction [J]. IEEE Access, 2019(7): 54472-54482. doi:  10.1109/ACCESS.2019.2913316
    [12]
    CHEN T T, WANG N B, HE M, et al. Reducing wrong labels for distantly supervised relation extraction with reinforcement learning [J]. IEEE Access, 2020, 99: 1-12. doi:  10.1109/ACCESS.2020.2990680
    [13]
    WU W Y, CHEN Y F, XU J N, et al. Attention-based convolutional neural for Chinese relation extraction [C]// Lecture Notes in Computer Science, vol 11221. Cham: Springer, 2018: 147-158.
    [14]
    WEN J, SUN X, REN X C, et al. Structure regularized neural network for entity relation classification for Chinese literature text [J]. Computing Research Repository, 2018, 171: 103-112. doi:  10.18653/v1/n18-2059
    [15]
    LI Z R, DING N, LIU Z Y, et al. Chinese relation extraction with multi-grained information and external linguistic knowledge [C]// Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 2019: 4377-4386.
    [16]
    张志昌, 周侗, 张瑞芳, 等. 融合双向GRU与注意力机制的医疗实体关系识别 [J]. 计算机工程, 2020, 46(6): 296-302.
    [17]
    MIKOLOV T, SUTSKEVER I, CHEN K, et al. Distributed representations of words and phrases and their compositionality [J]. Advances in Neural Information Processing Systems, 2013, 26: 3111-3119.
    [18]
    XU J J, WEN J, SUN X, et al. A discourse-level named entity recognition and relation extraction dataset for Chinese literature text [J]. IEEE Access, 2017(3): 25-32.
    [19]
    OLSON D L, DELEN D. Advanced Data Mining Techniques [M]. Berlin: Springer, 2008.
  • 加载中

Catalog

    通讯作者: 陈斌, bchen63@163.com
    • 1. 

      沈阳化工大学材料科学与工程学院 沈阳 110142

    1. 本站搜索
    2. 百度学术搜索
    3. 万方数据库搜索
    4. CNKI搜索

    Figures(2)  / Tables(3)

    Article views (177) PDF downloads(18) Cited by()
    Proportional views

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return