中文版 | English
题名

An extensive study on pre-trained models for program understanding and generation

作者
通讯作者Zhang,Yuqun
DOI
发表日期
2022-07-18
会议录名称
页码
39-51
摘要
Automatic program understanding and generation techniques could significantly advance the productivity of programmers and have been widely studied by academia and industry. Recently, the advent of pre-trained paradigm enlightens researchers to develop general-purpose pre-trained models which can be applied for a broad range of program understanding and generation tasks. Such pre-trained models, derived by self-supervised objectives on large unlabelled corpora, can be fine-tuned in downstream tasks (such as code search and code generation) with minimal adaptations. Although these pre-trained models claim superiority over the prior techniques, they seldom follow equivalent evaluation protocols, e.g., they are hardly evaluated on the identical benchmarks, tasks, or settings. Consequently, there is a pressing need for a comprehensive study of the pre-trained models on their effectiveness, versatility as well as the limitations to provide implications and guidance for the future development in this area. To this end, we first perform an extensive study of eight open-access pre-trained models over a large benchmark on seven representative code tasks to assess their reproducibility. We further compare the pre-trained models and domain-specific state-of-the-art techniques for validating pre-trained effectiveness. At last, we investigate the robustness of the pre-trained models by inspecting their performance variations under adversarial attacks. Through the study, we find that while we can in general replicate the original performance of the pre-trained models on their evaluated tasks and adopted benchmarks, subtle performance fluctuations can refute the findings in their original papers. Moreover, none of the existing pre-trained models can dominate over all other models. We also find that the pre-trained models can significantly outperform non-pre-trained state-of-the-art techniques in program understanding tasks. Furthermore, we perform the first study for natural language-programming language pre-trained model robustness via adversarial attacks and find that a simple random attack approach can easily fool the state-of-the-art pre-trained models and thus incur security issues. At last, we also provide multiple practical guidelines for advancing future research on pre-trained models for program understanding and generation.
关键词
学校署名
第一 ; 通讯
语种
英语
相关链接[Scopus记录]
收录类别
EI入藏号
20223512667335
EI主题词
Deep learning ; Natural language processing systems
EI分类号
Ergonomics and Human Factors Engineering:461.4 ; Data Processing and Image Processing:723.2
Scopus记录号
2-s2.0-85134872006
来源库
Scopus
引用统计
被引频次[WOS]:55
成果类型会议论文
条目标识符http://sustech.caswiz.com/handle/2SGJ60CL/401636
专题南方科技大学
工学院_斯发基斯可信自主研究院
作者单位
1.Southern University of Science and Technology,China
2.Southern University of Science and Technology,Hong Kong Polytechnic University,China
3.Kwai,China
4.Hong Kong Polytechnic University,China
5.University of Illinois at Urbana-Champaign,United States
6.Research Institute of Trustworthy Autonomous Systems,Shenzhen,China
第一作者单位南方科技大学
通讯作者单位南方科技大学
第一作者的第一单位南方科技大学
推荐引用方式
GB/T 7714
Zeng,Zhengran,Tan,Hanzhuo,Zhang,Haotian,et al. An extensive study on pre-trained models for program understanding and generation[C],2022:39-51.
条目包含的文件
条目无相关文件。
个性服务
原文链接
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
导出为Excel格式
导出为Csv格式
Altmetrics Score
谷歌学术
谷歌学术中相似的文章
[Zeng,Zhengran]的文章
[Tan,Hanzhuo]的文章
[Zhang,Haotian]的文章
百度学术
百度学术中相似的文章
[Zeng,Zhengran]的文章
[Tan,Hanzhuo]的文章
[Zhang,Haotian]的文章
必应学术
必应学术中相似的文章
[Zeng,Zhengran]的文章
[Tan,Hanzhuo]的文章
[Zhang,Haotian]的文章
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
[发表评论/异议/意见]
暂无评论

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。