Paper 2024/1228
Automated Software Vulnerability Static Code Analysis Using Generative Pre-Trained Transformer Models
Abstract
Generative Pre-Trained Transformer models have been shown to be surprisingly effective at a variety of natural language processing tasks -- including generating computer code. However, in general GPT models have been shown to not be incredibly effective at handling specific computational tasks (such as evaluating mathematical functions).
In this study, we evaluate the effectiveness of open source GPT models, with no fine-tuning, and with context introduced by the langchain and localGPT Large Language Model (LLM) framework, for the task of automatic identification of the presence of vulnerable code syntax (specifically targeting C and C++ source code). This task is evaluated on a selection of
Metadata
- Available format(s)
-
PDF
- Category
- Applications
- Publication info
- Preprint.
- Keywords
- CWECommon Weakness EnumerationGPT modelGenerative Pre-Trained TransformerStatic Code analysisNIST SARD dataset
- Contact author(s)
- elijah pelofske @ protonmail com
- History
- 2024-08-02: approved
- 2024-07-31: received
- See all versions
- Short URL
- https://ia.cr/2024/1228
- License
-
CC BY
BibTeX
@misc{cryptoeprint:2024/1228, author = {Elijah Pelofske and Vincent Urias and Lorie M. Liebrock}, title = {Automated Software Vulnerability Static Code Analysis Using Generative Pre-Trained Transformer Models}, howpublished = {Cryptology {ePrint} Archive, Paper 2024/1228}, year = {2024}, url = {https://eprint.iacr.org/2024/1228} }