China Safety Science Journal ›› 2025, Vol. 35 ›› Issue (3): 204-211.doi: 10.16265/j.cnki.issn1003-3033.2025.03.0223

• Public safety • Previous Articles     Next Articles

Gas knowledge bidirectional encoder representations from transformers model based on knowledge injection

LIU Xiaoyu(), ZHUANG Yufeng**(), ZHAO Xinghao, WANG Kefan, ZHANG Guokai   

  1. School of Intelligent Engineering and Automation, Beijing University of Posts and Telecommunications, Beijing 100876, China
  • Received:2024-10-14 Revised:2024-12-18 Online:2025-03-28 Published:2025-09-28
  • Contact: ZHUANG Yufeng

Abstract:

In order to enhance emergency management in the field of gas pipeline networks, Gas-kBERT model was proposed. The model incorporated data from the gas pipeline network field expanded by Chat Generative Pre-Trained Transformer,(ChatGPT)and Chinese Gas Language Understanding Subject-Predicate-Object(CGLU-Spo) and related corpora were constructed in this field. By altering the model's masking (MASK) mechanism, domain knowledge was successfully injected into the model. Considering the professionalism and specificity of the gas pipeline network field, Gas-kBERT was pre-trained on various scales and contents of corpora and fine-tuned on named entity recognition and classification tasks within this field. Experimental results demonstrated that, compared to the general BERT model, Gas-kBERT exhibited significant performance improvements in F1-score in text mining tasks in the gas pipeline network field. Specifically, in the named entity recognition task, the F1-score was increased by 29.55%, and in the text classification task, the F1-score improvement reached up to 83.33%. This study proves that the Gas-kBERT model performs exceptionally well in text mining tasks in the gas pipeline network field.

Key words: gas pipeline networks, gas knowledge bidirectional encoder representations from transformers(Gas-kBERT)model, natural language processing(NLP), knowledge injection, bidirectional encoder representations from transformers (BERT)

CLC Number: