Multi-role Consensus through LLMs Discussions for Vulnerability Detection
Mao, Zhenyu, Li, Jialong, Jin, Dongming, Li, Munan, Tei, Kenji
–arXiv.org Artificial Intelligence
Abstract--Recent advancements in large language models tester receives the initial prompt detailing its role-setting, its (LLMs) have highlighted the potential for vulnerability detection, task, and the code segment to analyze. The tester is asked to a crucial component of software quality assurance. The response viewpoints from different roles in a typical software development is constrained by a maximum token limit, ensuring that the life-cycle, including both developers and testers. Preliminary evaluation of this approach indicates iterative output exchange in an attempt to reach a collectively a 13.48% increase in the precision rate, an 18.25% increase multi-perspective consensus inside the code review team. in the recall rate, and a 16.13% increase in the F1 score. The tester and the developer, equipped with their unique perspective and judgments, enter a dialectic interaction, aimed Keywords-large language models; vulnerability detection; at exploring and resolving different opinions on potential prompt engineering; software quality assurance vulnerabilities.
arXiv.org Artificial Intelligence
May-18-2024