[Review] Examining Zero-Shot Vulnerability Repair with Large Language Models
The paper tests the performance of LLM for program repair. The same topic as Automated Program Repair in the Era of Large Pre-trained Language Models. Differently, this paper focuses more on the details, whose program repair setting is much more complicated.
Some conclusions were drawn:
- LLMs can generate fixes to bugs.
- But for real-world settings, the performance is not enough.
Background:
- Security bugs are significant.
- LLMs are popular and has outstanding performance.
Implementation:
RQ1: Can off-the-shelf LLMs generate safe and functional code to fix security vulnerabilities?
RQ2: Does varying the amount of context in the comments of a prompt affect the LLM’s ability to suggest fixes?
RQ3: What are the challenges when using LLMs to fix vulnerabilities in the real world?
RQ4: How reliable are LLMs at generating repairs?
![[Review] Examining Zero-Shot Vulnerability Repair with Large Language Models](/blog/images/24/cover.jpg)
![[Review] Titan : Efficient Multi-target Directed Greybox Fuzzing](/blog/images/19/cover.jpg)
![[Review] How IoT Re-using Threatens Your Sensitive Data: Exploring the User-Data Disposal in Used IoT Devices](/blog/images/13/cover.png)
![[Review] Assisting Static Analysis with Large Language Models: A ChatGPT Experiment](/blog/images/42/cover.png)
![[Review] Detecting Missed Security Operations Through Differential Checking of Object-based Similar Paths](/blog/images/41/cover.png)
![[Review] GPTScan: Detecting Logic Vulnerabilities in Smart Contracts by Combining GPT with Program Analysis](/blog/images/40/cover.png)
![[Review] MoonShine: Optimizing OS Fuzzer Seed Selection with Trace Distillation](/blog/images/39/cover.png)
![[Review] One Simple API Can Cause Hundreds of Bugs: An Analysis of Refcounting Bugs in All Modern Linux Kernels](/blog/images/38/cover.png)