Google’s Project Zero, a team of security analysts, has introduced a new framework aimed at enhancing automated vulnerability research using large language models.
Project Naptime uses AI to replicate the systematic methods used by human security researchers to reduce some of the pressure on the already strained workforce.
The initiative gets its name from its potential to allow human workers to “take regular naps” while AI handles complex vulnerability research tasks.
Google reveals details of Project Naptime
Google Project Zero’s Sergei Glazunov and Mark Brand, noted, “Naptime uses a specialised architecture to enhance an LLM’s ability to perform vulnerability research.”
Key components of the Naptime architecture include a Code Browser Tool which allows the AI agent to navigate the target codebase, similar to how engineers use Chromium Code Search; a Python Tool that enables running Python scripts in a sandboxed environments, a Debugger Tool that observes program behavior with different inputs; and a Reporter Tool that monitors the task progress and verifies success conditions.
Glazunov and Brand added: “Naptime enables an LLM to perform vulnerability research that closely mimics the iterative, hypothesis-driven approach of human security experts.”
In tests using the CyberSecEval 2 benchmark suite, released by rival tech company Meta, Naptime demonstrated significant improvements in identifying buffer overflow and advanced memory corruption flaws in C and C++ code.
Though in its early stages, Google’s Project Naptime marks a significant step forward in automated vulnerability research, potentially helping to reduce gaps left by traditional methods while addressing the ongoing skills shortage.