IDEAS home Printed from https://ideas.repec.org/a/gam/jsusta/v14y2022i19p12299-d927179.html
   My bibliography  Save this article

Graph Embedding-Based Domain-Specific Knowledge Graph Expansion Using Research Literature Summary

Author

Listed:
  • Junho Choi

    (Division of Undeclared Majors, Chosun University, Gwangju 61452, Korea)

Abstract

Knowledge bases built in the knowledge processing field have a problem in that experts have to add rules or update them through modifications. To solve this problem, research has been conducted on knowledge graph expansion methods using deep learning technology, and in recent years, many studies have been conducted on methods of generating knowledge bases by embedding the knowledge graph’s triple information in a continuous vector space. In this paper, using a research literature summary, we propose a domain-specific knowledge graph expansion method based on graph embedding. To this end, we perform pre-processing and process and text summarization with the collected research literature data. Furthermore, we propose a method of generating a knowledge graph by extracting the entity and relation information and a method of expanding the knowledge graph using web data. To this end, we summarize research literature using the Bidirectional Encoder Representations from Transformers for Summarization (BERTSUM) model based on domain-specific research literature data and design a Research-BERT (RE-BERT) model that extracts entities and relation information, which are components of the knowledge graph, from the summarized research literature. Moreover, we proposed a method of expanding related entities based on Google news after extracting related entities through the web for the entities in the generated knowledge graph. In the experiment, we measured the performance of summarizing research literature using the BERTSUM model and the accuracy of the knowledge graph relation extraction model. In the experiment of removing unnecessary sentences from the research literature text and summarizing them in key sentences, the result shows that the BERTSUM Classifier model’s ROUGE-1 precision is 57.86%. The knowledge graph extraction performance was measured using the mean reciprocal rank (MRR), mean rank (MR), and HIT@N rank-based evaluation metric. The knowledge graph extraction method using summarized text showed superior performance in terms of speed and knowledge graph quality.

Suggested Citation

  • Junho Choi, 2022. "Graph Embedding-Based Domain-Specific Knowledge Graph Expansion Using Research Literature Summary," Sustainability, MDPI, vol. 14(19), pages 1-15, September.
  • Handle: RePEc:gam:jsusta:v:14:y:2022:i:19:p:12299-:d:927179
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2071-1050/14/19/12299/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2071-1050/14/19/12299/
    Download Restriction: no
    ---><---

    References listed on IDEAS

    as
    1. Taejin Kim & Yeoil Yun & Namgyu Kim, 2021. "Deep Learning-Based Knowledge Graph Generation for COVID-19," Sustainability, MDPI, vol. 13(4), pages 1-19, February.
    2. Jongmo Kim & Kunyoung Kim & Mye Sohn & Gyudong Park, 2022. "Deep Model-Based Security-Aware Entity Alignment Method for Edge-Specific Knowledge Graphs," Sustainability, MDPI, vol. 14(14), pages 1-22, July.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Hind Bitar & Amal Babour & Fatema Nafa & Ohoud Alzamzami & Sarah Alismail, 2022. "Increasing Women’s Knowledge about HPV Using BERT Text Summarization: An Online Randomized Study," IJERPH, MDPI, vol. 19(13), pages 1-15, July.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jsusta:v:14:y:2022:i:19:p:12299-:d:927179. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.