Training Large Language Models to Follow System Prompt with Self-Supervised Fine-Tuning
Junyan Qiu1,2; Haitao Wang3; Yiping Yang2
2024-03
会议名称International Joint Conference on Neural Networks
会议日期2024-07
会议地点YOKOHAMA, JAPAN
出版者IEEE
摘要

In the realm of artificial intelligence, system prompts
stand as directives or requests aimed at guiding systems, such
as programming environments or AI models, to execute specific
tasks or operations. Typically positioned at the commencement
of input sequences in large language models, these prompts play
a pivotal role in shaping the model’s response and guiding its
interaction flow. However, a notable challenge emerges during
multi-turn dialogues, where these models gradually diverge from
adhering to the initial system prompt, leading to inconsistencies
in the dialogue. In this paper, we present a scalable framework
facilitating the adherence of language models to system prompts
through automated data construction. Our approach, termed
SELF-SUPERVISED SYSTEM PROMPT FINE-TUNING (S3FT), be-
gins by prompting a language model to modify real dialogue
responses to fit a specific system prompt, using stylized transla-
tion. Subsequently, we select a small sample of these responses
for human preference annotation. This annotated data is utilized
to train the language model to act as a discriminator, identi-
fying high-quality examples that are then employed in further
supervised fine-tuning. Experimental results on several datasets
demonstrate that applying our method to LlaMA2 and ChatGLM
promotes human preference rates by over 50%, and outperforms
ChatGPT and GPT4 by a consideratble margin. The source code
of our paper is available in S3FT-repo.

关键词large language models supervised fine-tuning instruct tuning stylized generation
学科门类工学::计算机科学与技术(可授工学、理学学位)
收录类别EI
是否为代表性论文
七大方向——子方向分类自然语言处理
国重实验室规划方向分类语音语言处理
是否有论文关联数据集需要存交
文献类型会议论文
条目标识符http://ir.ia.ac.cn/handle/173211/57413
专题综合信息系统研究中心_视知觉融合及其应用
通讯作者Junyan Qiu
作者单位1.University of Chinese Academy of Sciences
2.Institute of Automation, Chinese Academy of Sciences
3.Meituan
第一作者单位中国科学院自动化研究所
通讯作者单位中国科学院自动化研究所
推荐引用方式
GB/T 7714
Junyan Qiu,Haitao Wang,Yiping Yang. Training Large Language Models to Follow System Prompt with Self-Supervised Fine-Tuning[C]:IEEE,2024.
条目包含的文件 下载所有文件
文件名称/大小 文献类型 版本类型 开放类型 使用许可
a185-qiu final.pdf(1596KB)会议论文 开放获取CC BY-NC-SA浏览 下载
个性服务
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
谷歌学术
谷歌学术中相似的文章
[Junyan Qiu]的文章
[Haitao Wang]的文章
[Yiping Yang]的文章
百度学术
百度学术中相似的文章
[Junyan Qiu]的文章
[Haitao Wang]的文章
[Yiping Yang]的文章
必应学术
必应学术中相似的文章
[Junyan Qiu]的文章
[Haitao Wang]的文章
[Yiping Yang]的文章
相关权益政策
暂无数据
收藏/分享
文件名: a185-qiu final.pdf
格式: Adobe PDF
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。