oss-sec mailing list archives
Re: Coordinated Disclosure in the LLM Age
From: Greg Dahlman <dahlman () gmail com>
Date: Tue, 28 Apr 2026 14:09:36 -0600
As I have been struggling about the ethics of releasing a POC I discussed on this list (which just shared previously disclosed issues) let me provide some feedback. 1) While most model providers are following the common dark pattern of implicitly opting in non-enterprise users into data collection for training, the maximum acceptable embargo period for issues disclosed to these lists is 14 days, way shorter than the training period for the foundational models. I don't think that cached data is used across customers right now, it would seem too hard to avoid malicious prompts etc... to do so. In fact 90 days for an upstream disclosure is probably too short for that to be a *primary* concern. 2) Unless a bug happens to hit an exact match past the interpolation threshold, it is already well represented in the training data, but perhaps not in a pattern form that us humans will view as equal, so yes it will change how you triage. 3) While I personally have a bright line, where I won't even put client data on shared services, you should assume that any person who receives content will potentially leak it. 4) The AI tooling is probably higher risk as many of these projects (and traditional services) are passing the buck on security, note the below quote from cursor from [0]
Right now, the agent can read files without confirmation. That’s the
current behavior by design, and it’s described in the security docs. The read restriction only applies to files listed in .cursorignore. [0] https://forum.cursor.com/t/cursor-agents-should-be-restricted-from-access-of-files-outside-the-project-without-permission/149418 Security/friction tradeoff is always a challenge, but let me address the point below:
any vulnerability reported as a result of research using an LLM is
trivially discoverable by others, and give up trying to pretend there's any point to working it under embargo. There is nuance here, was it discovered by someone pumping your code into Qwen3.6-35B-A3B on a Mac Mini, or was it discovered by Anthropic burning tens of thousands of dollars worth of tokens? There is an almost uncountable amount of vulnerabilities, simply because the threat landscape has changed, from the MIT AI labs needs, the rainbow books, legacy code, indifference, etc.... While LLMs change the cost of discovery and reporting, they unfortunately may actually exacerbate the problem of mitigation. The corpus is just full of code that people are just glad it works, and LLMs have dramatically improved the percentage of code they produce that is functionally 'correct', they have not dramatically improved the security of the code they produce at the same rate. Even the most simplistic synthetic benchmarks on LLM's ability to produce correct *and* secure code at the same time are sitting at the coin flip range today. Meaning the efficiency gains in reporting vulnerabilities will be asymmetric with the ability to address them. Any change in embargo timelines needs to respect that asymmetry, and forgoing an embargo doesn't help lighten the load or get fixes in place; it will just add context switches and reduce velocity more. Those disruptive callouts need to be limited to the fire-drills where there is actually a fire because they are super expensive. Hopefully others have better feedback, but understand that with current management practices at many peoples day jobs, you may need to approach people in private for frank feedback. Greg On Tue, Apr 28, 2026 at 8:58 AM Jeremy Stanley <fungi () yuggoth org> wrote:
As I'm sure is the case for everyone, the projects I work in are under a seemingly unending deluge of vulnerability reports from researchers using LLMs to mine for security gold in our software. At the same time, we see maintainers on our projects relying on LLM-oriented tools to develop fixes for vulnerabilities and compose prose for advisories. While I take a moment to catch my breath, this new Bizarro World we're all living in has gotten me thinking about the risks of public LLM services to embargoed vulnerability handling workflows and traditional coordinated disclosure. The operators of these LLM services are known to feed prompts and results back into their training data, presumably making it faster and easier for the same information to be found later by other users of the same service. Would keeping embargoes short help to mitigate related risks of parallel rediscovery or outright disclosure to other LLM users? It seems to me that there must be some inherent lag in this process, but how much? I'm sorely tempted, both due to the increased volume and the risk of premature disclosure, to just assume that any vulnerability reported as a result of research using an LLM is trivially discoverable by others, and give up trying to pretend there's any point to working it under embargo. Similarly, it makes sense to me that patch development and descriptive prose shouldn't be produced with LLM assistance for any vulnerability that is being worked under an embargo. I can't be the only one whose been pondering this... what positions have the rest of you taken? [P.S. No LLMs were exploited in the making of this message.] -- Jeremy Stanley
Current thread:
- Coordinated Disclosure in the LLM Age Jeremy Stanley (Apr 28)
- Re: Coordinated Disclosure in the LLM Age Greg Dahlman (Apr 28)
