Google’s Big Sleep AI model just found a zero-day vulnerability in the wild — but don’t hold your breath for game-changing AI bug hunting tools any time soon
Google clarified it was the first undiscovered memory safety bug to be flagged by an AI agent, touting this as a significant step in using AI for vulnerability research


Google has claimed a vulnerability flagged by its Big Sleep AI model represents the first time an AI tool has found an unknown bug in the wild.
Google clarified that this is the first time such a system has detected a memory-safety bug, acknowledging other AI tools have discovered different types of vulnerabilities before.
“We believe this is the first public example of an AI agent finding a previously unknown exploitable memory-safety issue in widely used real-world software,” stated the blog post from the Big Sleep Team.
The vulnerability was described as an exploitable stack buffer underflow flaw in SQLite, the most widely deployed open source database engine. The flaw could have allowed an attacker to intentionally crash or execute arbitrary code to subvert existing security software on the system.
The team behind the AI model said once they discovered the flaw, they reported it to the developers in early October, who fixed it on the same day. Researchers noted the issue was fixed before it appeared in an official release, meaning no SQLite users were not impacted.
Google began using SQLite to test Big Sleep’s bug-hunting capabilities after seeing AI security research organization Team Atlanta use their Atlantis cyber reasoning system (CRS) to find a null pointer dereference at the DARPA AIxCC event.
This inspired the Big Sleep team to see if they could find a more serious vulnerability using their newly developed LLM.
Get the ITPro daily newsletter
Sign up today and you will receive a free copy of our Future Focus 2025 report - the leading guidance on AI, cybersecurity and other IT challenges as per 700+ senior executives
The researchers noted that exploiting the vulnerability was not trivial and thus would have proved difficult for threat actors to leverage into a successful attack, but demonstrates Big Sleep’s bug hunting prowess nonetheless.
Flaw flagged by Big Sleep slips through the cracks of traditional fuzzing methods
Big Sleep was a collaborative effort developed by Google’s zero-day hunting team Project Zero and its DeepMind AI research lab.
The tool is an evolution of earlier versions of Google’s framework for LLM-assisted vulnerability research known as Project Naptime, which was announced in June 2024.
Project Naptime was launched to evaluate the offensive security capabilities of LLMs, leveraging the rapidly improving code-comprehension of these models to “reproduce the systematic approach of a human security researcher when identifying and demonstrating security vulnerabilities”.
The post noted that fuzzing, the traditional approach to testing software for vulnerabilities by feeding it invalid or unexpected inputs, has limitations to find some flaws.
“A key motivating factor for Naptime and now for Big Sleep has been the continued in-the-wild discovery of exploits for variants of previously found and patched vulnerabilities, developers on the Big Sleep team said.
“As this trend continues, it's clear that fuzzing is not succeeding at catching such variants, and that for attackers, manual variant analysis is a cost-effective approach.”
In this case, the researchers reported that in their attempt to rediscover the flaw through fuzzing, it was not able to find the memory safety bug flagged by Big Sleep.
The post cautioned, however, that the results of the testing were “highly experimental”, and that the Big Sleep development team still hold that a target-specific fuzzer would be at least as effective at finding similar flaws.

Solomon Klappholz is a former staff writer for ITPro and ChannelPro. He has experience writing about the technologies that facilitate industrial manufacturing, which led to him developing a particular interest in cybersecurity, IT regulation, industrial infrastructure applications, and machine learning.
-
Should AI PCs be part of your next hardware refresh?
AI PCs are fast becoming a business staple and a surefire way to future-proof your business
By Bobby Hellard
-
Westcon-Comstor and Vectra AI launch brace of new channel initiatives
News Westcon-Comstor and Vectra AI have announced the launch of two new channel growth initiatives focused on the managed security service provider (MSSP) space and AWS Marketplace.
By Daniel Todd
-
Foreign AI model launches may have improved trust in US AI developers, says Mandiant CTO – as he warns Chinese cyber attacks are at an “unprecedented level”
News Concerns about enterprise AI deployments have faded due to greater understanding of the technology and negative examples in the international community, according to Mandiant CTO Charles Carmakal.
By Rory Bathgate
-
Adopting more security tools doesn't keep you safe, it just overloads your teams
News Security tool sprawl makes it harder to manage environments and overwhelms teams
By Ross Kelly
-
Australia and Google turn to AI to protect critical infrastructure
News Australia's CSIRO partners with Google to develop homegrown AI security tools for infrastructure
By Nicole Kobie
-
Google Workspace just got a slew of new zero trust features to help supercharge user security – here's what you need to know
News New Zscaler integrations across Chrome Enterprise, Google Workspace, and Google Security Operations aim to enhance enterprise security and access
By Daniel Todd
-
Google says Microsoft can’t be trusted after email security blunders
News Google has fired a broadside at Microsoft amid concerns over the tech giant's repeated security blunders
By Solomon Klappholz
-
Google forced to delete billions of incognito browsing records after privacy controversy
News Google has agreed to delete data it gained improperly through its private browsing function
By George Fitzmaurice
-
Google spent $10 million on bug bounty payouts last year — here's what flaws researchers uncovered
News Google’s Vulnerability program paid rewards to 600 researchers in 2023, with Android flaws earning a third of the total
By Steve Ranger
-
Magika, Google's new AI security tool, helps users identify malware at rapid speed - and it's free to access on GitHub
News Google hopes its Magika AI security tool will rapidly accelerate malicious file identification, but there are some limitations, according to industry experts
By George Fitzmaurice