With the continuous development of artificial intelligence and natural language processing technologies, traditional retrieval-augmented generation (RAG) techniques face numerous challenges in document answer precision and similarity measurement. This study, set against the backdrop of the shipping industry, combines top-down and bottom-up schema design strategies to achieve precise and flexible knowledge representation. The research adopts a semi-structured approach, innovatively constructing an adaptive schema generation mechanism based on reinforcement learning, which models the knowledge graph construction process as a Markov decision process. This method begins with general concepts, defining foundational industry concepts, and then delves into abstracting core concepts specific to the maritime domain through an adaptive pattern generation mechanism that dynamically adjusts the knowledge structure. Specifically, the study designs a four-layer knowledge construction framework, including the data layer, modeling layer, technology layer, and application layer. It draws on a mutual indexing strategy, integrating large language models and traditional information extraction techniques. By leveraging self-attention mechanisms and graph attention networks, it efficiently extracts semantic relationships. The introduction of logic-form-driven solvers and symbolic decomposition techniques for reasoning significantly enhances the model’s ability to understand complex semantic relationships. Additionally, the use of open information extraction and knowledge alignment techniques further improves the efficiency and accuracy of information retrieval. Experimental results demonstrate that the proposed method not only achieves significant performance improvements in knowledge graph retrieval within the shipping domain but also holds important theoretical innovation and practical application value.
References
[1]
Chen, J., Zhang, X., Xu, L. and Xu, J. (2024) Trends of Digitalization, Intelligence and Greening of Global Shipping Industry Based on CiteSpace Knowledge Graph. Ocean&CoastalManagement, 255, Article ID: 107206. https://doi.org/10.1016/j.ocecoaman.2024.107206
[2]
Wan, H., Fu, S., Zhang, M. and Xiao, Y. (2023) A Semantic Network Method for the Identification of Ship’s Illegal Behaviors Using Knowledge Graphs: A Case Study on Fake Ship License Plates. JournalofMarineScienceandEngineering, 11, Article 1906. https://doi.org/10.3390/jmse11101906
[3]
Zhang, Y., Xu, R., Lu, W., Mayer, W., Ning, D., Duan, Y., et al. (2023) Multi-Modal Spatio-Temporal Knowledge Graph of Ship Management. AppliedSciences, 13, Article 9393. https://doi.org/10.3390/app13169393
[4]
Li, Y., Liu, X., Wang, Z., Mei, Q., Xie, W., Yang, Y., et al. (2024) Construction of a Large-Scale Maritime Element Semantic Schema Based on Knowledge Graph Models for Unmanned Automated Decision-making. FrontiersinMarineScience, 11, Article 1390931. https://doi.org/10.3389/fmars.2024.1390931
[5]
Zhang, Z.Y., et al. (2019) ERNIE: Enhanced Language Representation with Informative Entities. arXiv: 1905.07129.
[6]
Liu, W., Zhou, P., Zhao, Z., Wang, Z., Ju, Q., Deng, H., et al. (2020) K-BERT: Enabling Language Representation with Knowledge Graph. ProceedingsoftheAAAIConferenceonArtificialIntelligence, 34, 2901-2908. https://doi.org/10.1609/aaai.v34i03.5681
[7]
Sun, T., Shao, Y., Qiu, X., Guo, Q., Hu, Y., Huang, X., et al. (2020) CoLAKE: Contextualized Language and Knowledge Embedding. Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, December 2020, 3660-3670. https://doi.org/10.18653/v1/2020.coling-main.327
[8]
Xiong, W.H., Du, J.F., Wang, W.Y. and Stoyanov, V. (2019) Pretrained Encyclopedia: Weakly Supervised Knowledge-Pretrained Language Model. arXiv: 1912.09637.
[9]
Yao, L., Mao, C.S. and Luo, Y. (2019) KG-BERT: BERT for Knowledge Graph Completion. arXiv: 1909.03193.
[10]
Yu, D., Zhu, C., Yang, Y. and Zeng, M. (2022) JAKET: Joint Pre-Training of Knowledge Graph and Language Understanding. ProceedingsoftheAAAIConferenceonArtificialIntelligence, 36, 11630-11638. https://doi.org/10.1609/aaai.v36i10.21417
[11]
Wang, R., Tang, D., Duan, N., Wei, Z., Huang, X., Ji, J., et al. (2021) K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters. Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, August 2021, 1405-1418. https://doi.org/10.18653/v1/2021.findings-acl.121
[12]
Zhang, H.Y., et al. (2023) Research on Question Answering System on Joint of Knowledge Graph and Large Language Models. Journal of Frontiers of Computer Science & Technology, 17, 2377. https://doi.org/10.3778/j.issn.1673-9418.2308070
[13]
Peters, M.E., et al. (2019) Knowledge Enhanced Contextual Word Representations. arXiv: 1909.04164.
[14]
He, B., Zhou, D., Xiao, J., Jiang, X., Liu, Q., Yuan, N.J., et al. (2020) BERT-MK: Integrating Graph Contextualized Knowledge into Pre-Trained Language Models. Findings of the Association for Computational Linguistics: EMNLP 2020, November 2020, 2281-2290. https://doi.org/10.18653/v1/2020.findings-emnlp.207
[15]
Yang, A., Wang, Q., Liu, J., Liu, K., Lyu, Y., Wu, H., et al. (2019) Enhancing Pre-Trained Language Representations with Rich Knowledge for Machine Reading Comprehension. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, July 2019, 2346-2357. https://doi.org/10.18653/v1/p19-1226
[16]
Huang, Z.H., et al. (2015) Bidirectional LSTM-CRF Models for Sequence Tagging. arXiv: 1508.01991.
[17]
Vaswani, A., et al. (2017) Attention is All you Need. arXiv: 1706.03762.
[18]
Mao, A.Q., Mohri, M. and Zhong, Y.T. (2023) Cross-Entropy Loss Functions: Theoretical Analysis and Applications. arXiv: 2304.07288.
[19]
Degris, T., White, M. and Sutton, R.S. (2012) Linear Off-Policy Actor-Critic. arXiv: 1205.4839.
[20]
Liang, L., Sun, M., Gui, Z., et al. (2024) KAG: Boosting LLMs in Professional Domains via Knowledge Augmented Generation. arXiv: 2409.13731.