Publishing Service

Polishing & Checking

Frontiers of Information Technology & Electronic Engineering

ISSN 2095-9184 (print), ISSN 2095-9230 (online)

Large language model-enhanced probabilistic modeling for effective static analysis alarms

Abstract: Static analysis presents significant challenges in alarm handling, where probabilistic models and alarm prioritization are essential methods for addressing these issues. These models prioritize alarms based on user feedback, thereby alleviating the burden on users to manually inspect alarms. However, they often encounter limitations related to efficiency and issues such as false generalization. While learning-based approaches have demonstrated promise, they typically incur high training costs and are constrained by the predefined structures of existing models. Moreover, the integration of large language models (LLMs) in static analysis has yet to reach its full potential, often resulting in lower accuracy rates in vulnerability identification. To tackle these challenges, we introduce BinLLM, a novel framework that harnesses the generalization capabilities of LLMs to enhance alarm probability models through rule learning. Our approach integrates LLM-derived abstract rules into the probabilistic model, using alarm paths and critical statements from static analysis. This integration enhances the model's reasoning capabilities, improving its effectiveness in prioritizing genuine bugs while mitigating false generalizations. We evaluated BinLLM on a suite of C programs and observed 40.1% and 9.4% reduction in the number of checks required for alarm verification compared to two state-of-the-art baselines, Bingo and BayeSmith, respectively, underscoring the potential of combining LLMs with static analysis to improve alarm management.

Key words: Static analysis; Bayesian inference; Large language models (LLMs); Alarm ranking

Chinese Summary  <0> 大语言模型增强的静态分析警报概率模型

潘鑫龙1,2,李建华1,2,周志洪1,2,李高磊1,2,陈秀真1,2,马进1,2,伍军1,2,张全海1,2
1上海交通大学计算机学院网络安全技术研究院,中国上海市,200240
2上海市信息安全综合管理技术研究重点实验室,中国上海市,200240
摘要:静态分析在警报处理方面面临诸多挑战,其中概率模型与警报优先级排序是缓解用户手动检查负担的关键方法。这些模型通常依赖用户反馈对警报进行排序,从而提升处理效率。然而,现有方法常受限于效率低下及泛化能力不足等问题。尽管基于学习的方法已展现一定潜力,但其通常伴随着高昂的训练代价,并受预定义模型结构的制约。此外,大语言模型(LLM)在静态分析中的集成尚未充分发挥其潜力,导致漏洞识别准确率偏低。为应对上述问题,本文提出一种新型框架—BinLLM,该框架利用LLM的泛化能力,通过规则学习提升警报概率模型的性能。我们的方法将LLM生成的抽象规则引入概率模型,结合静态分析中的警报路径与关键语句,从而增强模型推理能力,有效提高真实漏洞的识别率,并缓解泛化错误问题。在一组C程序的实验评估中,BinLLM在警报验证所需检查数量上,较两个先进基线方法Bingo和BayeSmith分别减少40.1%与9.4%,充分体现了LLM与静态分析的结合在提升警报管理方面的应用潜力。

关键词组:静态分析;贝叶斯推理;大语言模型;警报排序


Share this article to: More

Go to Contents

References:

<Show All>

Open peer comments: Debate/Discuss/Question/Opinion

<1>

Please provide your name, email address and a comment





DOI:

10.1631/FITEE.2500038

CLC number:

TP311.53;TP183

Download Full Text:

Click Here

Downloaded:

760

Download summary:

<Click Here> 

Downloaded:

24

Clicked:

709

Cited:

0

On-line Access:

2025-11-17

Received:

2025-01-16

Revision Accepted:

2025-11-18

Crosschecked:

2025-04-14

Journal of Zhejiang University-SCIENCE, 38 Zheda Road, Hangzhou 310027, China
Tel: +86-571-87952276; Fax: +86-571-87952331; E-mail: jzus@zju.edu.cn
Copyright © 2000~ Journal of Zhejiang University-SCIENCE