Author
Listed:
- Xin Tang
(Science and Technology on Information Systems Engineering Laboratory, National University of Defense Technology, Changsha 410073, China)
- Kunjia Liu
(Laboratory for Big Data and Decision, National University of Defense Technology, Changsha 410073, China)
- Hao Xu
(Laboratory for Big Data and Decision, National University of Defense Technology, Changsha 410073, China)
- Weidong Xiao
(Science and Technology on Information Systems Engineering Laboratory, National University of Defense Technology, Changsha 410073, China)
- Zhen Tan
(Science and Technology on Information Systems Engineering Laboratory, National University of Defense Technology, Changsha 410073, China)
Abstract
Prerequisite chains are crucial to acquiring new knowledge efficiently. Many studies have been devoted to automatically identifying the prerequisite relationships between concepts from educational data. Though effective to some extent, these methods have neglected two key factors: most works have failed to utilize domain-related knowledge to enhance pre-trained language models, thus making the textual representation of concepts less effective; they also ignore the fusion of semantic information and structural information formed by existing prerequisites. We propose a two-stage concept prerequisite learning model (TCPL), to integrate the above factors. In the first stage, we designed two continual pre-training tasks for domain-adaptive and task-specific enhancement, to obtain better textual representation. In the second stage, to leverage the complementary effects of the semantic and structural information, we optimized the encoder of the resource–concept graph and the pre-trained language model simultaneously, with hinge loss as an auxiliary training objective. Extensive experiments conducted on three public datasets demonstrated the effectiveness of the proposed approach. Our proposed model improved by 7.9%, 6.7%, 5.6%, and 8.4% on ACC, F1, AP, and AUC on average, compared to the state-of-the-art methods.
Suggested Citation
Xin Tang & Kunjia Liu & Hao Xu & Weidong Xiao & Zhen Tan, 2023.
"Continual Pre-Training of Language Models for Concept Prerequisite Learning with Graph Neural Networks,"
Mathematics, MDPI, vol. 11(12), pages 1-16, June.
Handle:
RePEc:gam:jmathe:v:11:y:2023:i:12:p:2780-:d:1175190
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:11:y:2023:i:12:p:2780-:d:1175190. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.