全部 标题 作者
关键词 摘要

OALib Journal期刊
ISSN: 2333-9721
费用:99美元

查看量下载量

相关文章

更多...

Exploring the Taxonomy of Survey Papers on Large Language Models Using Classical Machine Learning

DOI: 10.4236/jilsa.2025.172006, PP. 68-76

Keywords: Large Language Models (LLMs), Taxonomy, Multimodal Models

Full-Text   Cite this paper   Add to My Lib

Abstract:

The rapid advancements in large language models (LLMs) have led to an exponential increase in survey papers, making it challenging to systematically track and analyze their evolving taxonomy. This study employs graph representation learning combined with classical machine learning techniques to model and interpret the structural evolution of LLM-related survey papers. By constructing attributed graphs that capture topic distributions and interconnections, we provide a data-driven framework to explore research trends in this domain. A dataset of 241 survey papers published between July 2021 and January 2024 is analyzed to identify thematic developments and interdisciplinary relationships. The results highlight key areas of specialization, including the emergence of prompting science, multimodal models, and domain-specific applications in finance, education, and law. Co-occurrence analysis of survey topics reveals strong interconnections between core LLM research and fields such as software engineering, hardware architecture, and evaluation methodologies. These findings demonstrate the increasing specialization of LLM research and its growing integration across multiple disciplines. By leveraging graph-based methodologies, this study offers a structured approach to understanding the LLM survey landscape, facilitating efficient navigation of existing literature and identification of emerging research directions. The insights presented contribute to a more comprehensive understanding of the field’s trajectory, assisting researchers and practitioners in engaging with the latest developments in LLM research.

References

[1]  He, K., Zhang, X., Ren, S. and Sun, J. (2016) Deep Residual Learning for Image Recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, 27-30 June 2016, 770-778.
https://doi.org/10.1109/cvpr.2016.90
[2]  Dosovitskiy, A. (2020) An Image Is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv: 2010.11929.
[3]  Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Ł. and Polosukhin, I. (2017) Attention Is All You Need. arXiv: 1706.03762.
[4]  Devlin, J., Chang, M.W., Lee, K. and Toutanova, K. (2018) BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. arXiv: 1810.04805.
[5]  Kipf, T.N. and Welling, M. (2016) Semi-Supervised Classification with Graph Convolutional Networks. arXiv: 1609.02907.
[6]  Zhuang, J. and Al Hasan, M. (2022) Defending Graph Convolutional Networks against Dynamic Graph Perturbations via Bayesian Self-Supervision. Proceedings of the AAAI Conference on Artificial Intelligence, 36, 4405-4413.
https://doi.org/10.1609/aaai.v36i4.20362
[7]  Radford, A., Narasimhan, K., Salimans, T. and Sutskever, I. (2018) Improving Language Understanding by Generative Pre-Training. OpenAI.
[8]  Radford, A., Wu, J., Child, R., Luan, D., Amodei, D. and Sutskever, I. (2019) Language Models Are Unsupervised Multitask Learners. OpenAI Blog, 1, 8, 9.
[9]  Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G. and Askell, A. (2020) Language Models Are Few-Shot Learners. Advances in Neural Information Processing Systems, 33, 1877-1901.
[10]  Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, F.L., Almeida, D., Altenschmidt, J., Altman, S. and Anadkat, S. (2023) GPT-4 Technical Report. arXiv: 2303.08774.
[11]  Bai, Y., Kadavath, S., Kundu, S., Askell, A., Kernion, J., Jones, A., Chen, A., Goldie, A., Mirhoseini, A. and McKinnon, C. (2022) Constitutional AI: Harmlessness from AI Feedback. arXiv: 2212.08073.
[12]  Team, G., Anil, R., Borgeaud, S., Wu, Y., Alayrac, J.B., Yu, J., Soricut, R., Schalkwyk, J., Dai, A.M. and Hauth, A. (2023) Gemini: A Family of Highly Capable Multimodal Models. arXiv: 2312.11805.
[13]  Zhuang, J. and Kennington, C. (2024) Understanding Survey Paper Taxonomy about Large Language Models via Graph Representation Learning. Proceedings of the Fourth Workshop on Scholarly Document Processing (SDP 2024), Bangkok, August 2024, 58-69.
https://doi.org/10.18653/v1/2024.sdp-1.6

Full-Text

Contact Us

service@oalib.com

QQ:3279437679

WhatsApp +8615387084133