Multi-role Consensus through LLMs Discussions for Vulnerability Detection

Mao, Zhenyu, Li, Jialong, Jin, Dongming, Li, Munan, Tei, Kenji

arXiv.org Artificial Intelligence 

Abstract--Recent advancements in large language models tester receives the initial prompt detailing its role-setting, its (LLMs) have highlighted the potential for vulnerability detection, task, and the code segment to analyze. The tester is asked to a crucial component of software quality assurance. The response viewpoints from different roles in a typical software development is constrained by a maximum token limit, ensuring that the life-cycle, including both developers and testers. Preliminary evaluation of this approach indicates iterative output exchange in an attempt to reach a collectively a 13.48% increase in the precision rate, an 18.25% increase multi-perspective consensus inside the code review team. in the recall rate, and a 16.13% increase in the F1 score. The tester and the developer, equipped with their unique perspective and judgments, enter a dialectic interaction, aimed Keywords-large language models; vulnerability detection; at exploring and resolving different opinions on potential prompt engineering; software quality assurance vulnerabilities.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found