In the field of equipment support, the method of generating equipment support sentence vectors based on word vectors is simple and effective, but it ignores the order and dependency relationships between words in the sentence, thus failing to capture the overall semantic information of the sentence. In contrast, using deep learning models (such as RNN, LSTM, Transformer, etc.) to directly generate sentence vectors can better capture the order and dependency relationships between words in the sentence, and thus better represent the overall semantic information of the sentence, avoiding the loss of information by simplifying the sentence to the average or concatenation of word vectors. To address the characteristics of equipment support, a method for training equipment support domain sentence vectors based on Bert-Prompt is proposed to improve the semantic understanding and representation capabilities of equipment failure texts. Specifically, the pre-trained BERT model is applied to sentence vector training, and the concept of prompt learning is combined. By designing effective Prompt sentence vector templates and the InfoNCE Loss function, the representation effect of equipment support sentence vectors is further improved. Based on BERT-Prompt, the training of equipment support domain sentence vectors is explored. This includes an overview of BERT sentence vector models, the development of sentence vector models, common BERT sentence vector model introductions, an introduction to Bert-Prompt, the main achievements and innovations of Bert-Prompt, its core ideas, common strategies and methods of Bert-Prompt, template-based Prompt sentence vector representation, continuous Prompt templates, the InfoNCE Loss function, training and optimization processes. The experimental analysis section covers data preparation, evaluation metrics, experimental preparations, comparative experimental methods, and analysis of experimental results.
References
[1]
Duan, X., Zhang, Y. and Sun, Y. (2017) Research on Sentence Vector Representation and Similarity Calculation Method About Microblog Texts. Computer Engineering, 43, 143-148.
[2]
Mikolov, T., Chen, K., Corrado, G., et al. (2013) Efficient Estimation of Word Representations in Vector Space. arXiv: 1301.3781. https://doi.org/10.48550/arXiv.1301.3781
[3]
Le, Q. and Mikolov, T. (2014) Distributed Representations of Sentences and Documents. Proceedings of the 31st International Conference on International Conference on Machine Learning, Beijing, 21-26 June 2014, 1188-1196.
[4]
Kiros, R., Zhu, Y., Salakhutdinov, R.R., et al. (2015) Skip-Thought Vectors. arXiv: 1506.06726. https://doi.org/10.48550/arXiv.1506.06726
[5]
Conneau, A., Kiela, D., Schwenk, H., Barrault, L. and Bordes, A. (2017) Supervised Learning of Universal Sentence Representations from Natural Language Inference Data. Proceedingsofthe 2017 ConferenceonEmpiricalMethodsinNaturalLanguageProcessing, Copenhagen, 7-11 September 2017, 670-680. https://doi.org/10.18653/v1/d17-1070
[6]
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., et al. (2017) Attention Is All You Need. Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, 4-9 December 2017, 6000-6010.
[7]
Cer, D., Yang, Y., Kong, S., Hua, N., Limtiaco, N., St. John, R., etal. (2018) Universal Sentence Encoder for English. Proceedingsofthe 2018 ConferenceonEmpiricalMethodsinNaturalLanguageProcessing:SystemDemonstrations, Brussels, 31 October-4 November 2018, 169-174. https://doi.org/10.18653/v1/d18-2029
[8]
Devlin, J., Chang, M.W., Lee, K., et al. (2019) BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. Proceedings of NAACL-HLT 2019, Minneapolis, 2-7 June 2019, 4171-7186.
[9]
Liu, Y., Ott, M., Goyal, N., et al. (2019) RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv: 1907.11692. https://doi.org/10.48550/arXiv.1907.11692
[10]
Yang, Z., Dai, Z., Yang, Y., et al. (2020) XLNet: Generalized Autoregressive Pretraining for Language Understanding. arXiv: 1906.08237. https://doi.org/10.48550/arXiv.1906.08237
[11]
Lan, Z., Chen, M., Goodman, S., et al. (2020) ALBERT: A Lite BERT for Self-Supervised Learning of Language Representations. arXiv: 1909.11942. https://doi.org/10.48550/arXiv.1909.11942
[12]
Clark, K., Luong, M.T., Le, Q.V., et al. (2020) ELECTRA: Pre-Training Text Encoders as Discriminators Rather Than Generators. arXiv: 2003.10555. https://doi.org/10.48550/arXiv.2003.10555
[13]
Reimers, N. and Gurevych, I. (2019) Sentence-Bert: Sentence Embeddings Using Siamese Bert-networks. Proceedingsofthe 2019 ConferenceonEmpiricalMethodsinNaturalLanguageProcessingandthe 9thInternationalJointConferenceonNaturalLanguageProcessing (EMNLP-IJCNLP), Hong Kong, 3-7 November 2019, 3982-3992. https://doi.org/10.18653/v1/d19-1410
[14]
Su, J., Cao, J., Liu, W., et al. (2021) Whitening Sentence Representations for Better Semantics and Faster Retrieval. arXiv: 2103.15316. https://doi.org/10.48550/arXiv.2103.15316
[15]
Jiang, T., Jiao, J., Huang, S., Zhang, Z., Wang, D., Zhuang, F., etal. (2022) PromptBERT: Improving BERT Sentence Embeddings with Prompts. Proceedingsofthe 2022 ConferenceonEmpiricalMethodsinNaturalLanguageProcessing, Abu Dhabi, 7-11 December 2022, 8826-8837. https://doi.org/10.18653/v1/2022.emnlp-main.603
[16]
Brown, T.B., Mann, B., Ryder, N., et al. (2020) Language Models Are Few-Shot Learners. Proceedings of the 34th International Conference on Neural Information Processing System, Vancouver, 6-12 December 2020, 1877-1901.
[17]
Yu, B., Cai, X. and Wei, J. (2023) Few-Shot Text Classification Method Based on Prompt Learning. Journal of Computer Applications, 43, 2735-2740.
[18]
Li, N. (2022) Improved Sentence Embedding Based on BERT and Prompt-Learning. Shantou University.
[19]
Liu, X., Zheng, Y., Du, Z., et al. (2023) GPT Understands, Too. arXiv: 2103.10385. https://doi.org/10.48550/arXiv.2103.10385
[20]
Chen, X., Zhang, N., Xie, X., et al. (2022) KnowPrompt Knowledge-Aware Prompt-tuning with Synergistic Optimization for Relation Extraction. Proceedings of the ACM Web Conference 2022, Virtual, 25-29 April 2022, 2778-2788. https://doi.org/10.1145/3485447.3511998
[21]
Cho, K., van Merrienboer, B., Gulcehre, C., et al. (2014) Learning Phrase Representations Using RNNEncoder-Decoder for Statistical Machine Translation. Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, 25-29 October 2014, 1724-1734.
[22]
Gao, T., Yao, X. and Chen, D. (2021) SimCSE: Simple Contrastive Learning of Sentence Embeddings. Proceedingsofthe 2021 ConferenceonEmpiricalMethodsinNaturalLanguageProcessing, Punta Cana, 7-11 November 2021, 6894-6910. https://doi.org/10.18653/v1/2021.emnlp-main.552