节点文献
Llama2-70b模型的微调技术及其在材料领域的应用研究
A Study of the Fine-Tuning Technique of the Llama2-70b Model and Its Application in the Field of Materials
【摘要】 【目的】为降低大语言模型的使用门槛,促进大语言模型在学科领域的应用。本文系统介绍了Llama2-70b模型的微调过程及其在材料领域应用的流程。【方法】本研究利用DeepSpeed框架和无机材料合成路径的指令式数据集,采用LoRA微调技术对开源大模型Llama2-70b进行微调,并对模型的超参数进行了调优,从模型训练中的损失值和模型稳定性两个方面对调优效果进行了评估,最终确定了一组适合模型的超参数组合。【结果】通过对模型的训练和优化,最终获得了一个在稳定性和性能方面表现优异的材料合成大语言模型。【结论】该研究为大语言模型在学科领域的应用提供了宝贵的经验和方法,所训练的材料大语言模型为材料合成设计提供了有意义的参考和支持。
【Abstract】 [Objective] To lower the barriers of using large language models and promote their applications in different fields, this paper systematically introduces the fine-tuning process of the Llama2-70b model and its application procedure in the field of materials science. [Methods] This study utilized the DeepSpeed framework and an instruction data set of inorganic material synthesis pathways, and employed the LoRA fine-tuning technique to fine-tune the open-source Llama2-70b model.The model’s hyperparameters were optimized, and the tuning effects were evaluated based on the loss value during model training and the model’s stability. A suitable combination of hyperparameters was finally determined.[Results] Through the training and optimization of the model, a large language model for material synthesis that performs excellently in terms of stability and performance was obtained. [Conclusions] This research provides valuable experience and methods for the application of large language models in academic fields. The trained material language model offers meaningful reference and support for material synthesis design.
【Key words】 Llama2-70b Model; LoRA; Large Language model; material synthesis;
- 【文献出处】 数据与计算发展前沿(中英文) ,Frontiers of Data & Computing , 编辑部邮箱 ,2025年01期
- 【分类号】TB30;TP18
- 【下载频次】351