检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
作 者:秦小林[1,2] 古徐 李弟诚 徐海文 QIN Xiaolin;GU Xu;LI Dicheng;XU Haiwen(Chengdu Institute of Computer Application,Chinese Academy of Sciences,Chengdu Sichuan 610213,China;School of Computer Science and Technology,University of Chinese Academy of Sciences,Beijing 100049,China;Faculty of Science,Civil Aviation Flight University of China,Guanghan Sichuan 618307,China)
机构地区:[1]中国科学院成都计算机应用研究所,成都610213 [2]中国科学院大学计算机科学与技术学院,北京100049 [3]中国民用航空飞行学院理学院,四川广汉618307
出 处:《计算机应用》2025年第3期685-696,共12页journal of Computer Applications
基 金:国家重点研发计划项目(2023YFB3308601);四川省科技计划项目(2024NSFJQ0035,2024NSFSC0004);四川省委组织部人才专项。
摘 要:大语言模型(LLM)是由具有大量参数(通常数十亿个权重或更多)的人工神经网络组成的一类语言模型,使用自监督学习或半监督学习对大量未标记文本进行训练,是当前生成式人工智能(AI)技术的核心。与传统语言模型相比,LLM通过大量的算力、参数和数据支持,展现出更强的语言理解与生成能力,广泛应用于机器翻译、问答系统、对话生成等众多任务中并表现卓越。现有的综述大多侧重于LLM的理论架构与训练方法,对LLM的产业级应用实践及技术生态演进的系统性探讨仍显不足。因此,在介绍LLM的基础架构、训练技术及发展历程的基础上,分析当前通用的LLM关键技术和以LLM为底座的先进融合技术。通过归纳总结现有研究,进一步阐述LLM在实际应用中面临的挑战,包括数据偏差、模型幻觉和计算资源消耗等问题,并对LLM的持续发展趋势进行展望。Large Language Models(LLMs)are a class of language models composed of artificial neural networks with a vast number of parameters(typically billions of weights or more).They are trained on a large amount of unlabeled text using self-supervised or semi-supervised learning and are the core of current generative Artificial Intelligence(AI)technologies.Compared to traditional language models,LLMs demonstrate stronger language understanding and generation capabilities,supported by substantial computational power,extensive parameters,and large-scale data.They are widely applied in tasks such as machine translation,question answering systems,and dialogue generation with good performance.Most of the existing surveys focus on the theoretical construction and training techniques of LLMs,while systematic exploration of LLMs’industry-level application practices and evolution of the technological ecosystem remains insufficient.Therefore,based on introducing the foundational architecture,training techniques,and development history of LLMs,the current general key technologies in LLMs and advanced integration technologies with LLMs bases were analyzed.Then,by summarizing the existing research,challenges faced by LLMs in practical applications were further elaborated,including problems such as data bias,model hallucination,and computational resource consumption,and an outlook was provided on the ongoing development trends of LLMs.
关 键 词:大语言模型 智能体 自然语言处理 检索增强生成 模型幻觉
分 类 号:TP182[自动化与计算机技术—控制理论与控制工程]
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:18.191.163.22