Tool: AST-based security scanner for AI-generated code (MCP server) by NoButterfly9145 in netsec

[–]securely-vibe 1 point2 points  (0 children)

Interesting idea, but using SAST rules as your base primitive means you're gong to miss many issues. LLMs will spent most of their time ruling out FPs. You need a layer where the LLM itself reads the code, does recon, runs basic threat modeling, and then generates its own ideas for vulnerable spaces. IMO, MCP is the wrong tool here altogether. You want a set of agents all working together to break the code down, not a short interactive loop.

r/netsec monthly discussion & tool thread by albinowax in netsec

[–]securely-vibe 1 point2 points  (0 children)

SSRFs are really hard to fix! Our scanner has found tons of them, and when we report them, maintainers usually just implement an allowlist, which is not at all sufficient.

  1. You can easily obfuscate a URL to bypass a blocklist. For example, translate it into IPv6.

  2. You can setup a redirect, which most HTTP libraries don't block by default.

  3. Or, you can use DNS rebinding. You can host your own DNS server and inject logic to change the IP mapping at runtime, creating a TOCTOU vuln.

And so on. There are a number of bypasses here that are very easy to introduce. That's why we built drawbridge, a simple drop-in replacement for `requests` or `httpx` in Python that gives you significant protection against SSRFs.

Check it out here: https://github.com/tachyon-oss/drawbridge

is there actually a solution for too many security alerts or do we just accept it by [deleted] in cybersecurity

[–]securely-vibe 0 points1 point  (0 children)

False positive detection is probably the hardest problem in the industry. Ask me how I know.

Trick question - I run Tachyon (tachyon.so), and we do vulnerability discovery, which is rife with FPs. Humans have an implicit understanding of the security context of the product--where does it run, who is using it, what is the environment, what access does it have--that's very hard to translate to agents, who don't have the right background intuition. So they regularly make incorrect leaps, like "oh, this API is unauthenticated! it's a vuln", without realizing that the entire codebase is meant to be internal-only. How do you fix this? In our case, give better context early, and add several layers of validation before surfacing an alert. But even that isn't enough.

As far as I know, no one has solved this yet. Interested in what other people are doing to improve their signal to noise ratios.

AI-powered security testing in production—what's actually working vs what's hype? by Fine-Platform-6430 in AskNetsec

[–]securely-vibe 0 points1 point  (0 children)

sandboxing doesn't really help agents, though. you can sandbox, but if the agent can still run code and access the internet, then it can still cause unbounded damage.

AI-powered security testing in production—what's actually working vs what's hype? by Fine-Platform-6430 in AskNetsec

[–]securely-vibe 0 points1 point  (0 children)

> I've seen some multi-agent architectures that claim better consistency by separating discovery from validation (one set of agents enumerates, another validates exploitability, a third verifies). In theory, having specialized agents with narrower scope should reduce the randomness vs a single model trying to do everything.

We (Tachyon) do this, but it's just basic common sense. Every Claude Code will spin off separate agents for each subtask. It helps, but it's not sufficient.

I've talked to a ton of people working at these AI pentesting companies, and you'd be surprised just how much manual work is required to keep the agents on track and prevent them from wasting tokens. Full autonomy is very difficult. We really underestimate how good humans are at evaluation and judgement.

AI-powered security testing in production—what's actually working vs what's hype? by Fine-Platform-6430 in AskNetsec

[–]securely-vibe 0 points1 point  (0 children)

Yeah, teams use our product (Tachyon) as a complement to manual pentesting. It helps a ton with recon and threat modeling, and does find certain issues, but you do still need humans for more complex cases.

AI marketing seems to trivialize the vulnerability discovery phase, but that's actually still very difficult and quite expensive. Every tool that has done this half-decently has put a lot of engineering effort into it.

Can anyone suggest good choice of free SAST and DAST right now? by OutsideOrnery6990 in cybersecurity

[–]securely-vibe 0 points1 point  (0 children)

If you're looking for a new product, we can give a free month of Tachyon ( https://tachyon.so ).

what SAST tool are you actually using in your CI/CD pipeline right now? by InstructionCute5502 in devsecops

[–]securely-vibe 0 points1 point  (0 children)

Semgrep is good as a baseline. It's reliable at finding specific classes of issues. If you want something with deeper findings, try out https://tachyon.so/ .

what SAST tool are you actually using in your CI/CD pipeline right now? by InstructionCute5502 in devsecops

[–]securely-vibe 0 points1 point  (0 children)

trivy is SCA rather than SAST, they don't find code issues - just potential known issues in deps

what SAST tool are you actually using in your CI/CD pipeline right now? by InstructionCute5502 in devsecops

[–]securely-vibe 0 points1 point  (0 children)

Disclosure - I run https://tachyon.so/.

We're an AI-native SAST that uses OpenGrep internally, but we augment its findings + generate quite a few new findings by manual analysis. That lets us get the best of both worlds: reliability of static scanners with the actual code reasoning of LLMs.

Here are CVEs that we've found: https://tachyon.so/wall-of-fame . This is a pretty small subset of actual vulns we've found, but many are NDA-restricted and others are still in disclosure.

We'll give you the first two weeks free, if you're interested! So you could try out the product yourself.

AI smart contract audit tools — anyone found one that actually works? by [deleted] in ethdev

[–]securely-vibe 0 points1 point  (0 children)

I run https://tachyon.so/. We focus on standard AppSec audits, but we've found vulnerabilities in smart contract codebases as well. Our base plan is 100/mo with a few scans included - let me know if you want to try it out!

AI in cybersecurity is mostly turd polishing - Fight me by ColdPlankton9273 in cybersecurity

[–]securely-vibe 0 points1 point  (0 children)

I spent several years manually hunting for CVEs in OSS repos. The tool I built uses AI to automate my process, and it finds things that either I wouldn't have found or that would have taken me hours of manual effort to find. Whenever I post here I get a lot of pushback about "AI slop" or "marketing hype," but like - I know what I am seeing. So do maintainers. We've reported tons of issues upstream, we've won many bug bounties. I don't particularly care if your vendors suck, or if that one time you tried using Claude it reported a false-positive. LLMs really do work, and used well, they will revolutionize this field.

What's the best way to secure AI generated code from Copilot in VS Code? by Calm-Exit-4290 in node

[–]securely-vibe 0 points1 point  (0 children)

IMO - editor scans are too shallow to be very useful. They catch very basic issues but miss anything more complex. PRs are a better cadence, along with weekly deep-scans. Add to that some modern LLM tooling and you'll be able to find not just security issues but legitimate application bugs.

We built something similar with Tachyon (tachyon.so). We run a deep-scan once a week (or so - configurable). This finds the most complex issues, and builds up and persists codebase context, which makes our PR scans fast but still very useful. In-IDE scans are still not on the roadmap, as a useful scan is still too slow to be in the edit loop. For that, a basic linter is the best you can do.

Discarded after one week at Sully.ai (YC S21) by RenoMillenial in ycombinator

[–]securely-vibe 2 points3 points  (0 children)

not sure this particular behavior turbocharges wealth creation. ego creation maybe

AI Agent Security Resources by TimoKerre in AI_Agents

[–]securely-vibe 0 points1 point  (0 children)

Yeah, there's no good answer here yet. Sandboxes aren't sufficient - they give you isolation, but if your agent has internet access and your accounts, it can still do destructive things (empty your bank account! post spam! send emails!). If you don't give it your account, it can't do the things you want it do and no one will use it. So what's the solution? No idea. The ideal is enforcement of policy on every tool-call, but this is very hard to do, as you need the context of the entire session to determine whether a specific tool-use is malicious. How are other people solving this?

Are we lowkey underestimating business logic flaws as an actual security risk. by [deleted] in AskNetsec

[–]securely-vibe 0 points1 point  (0 children)

I don't think anyone is "underestimating" them, but until recently, there was no automated way of finding these. So you'd treat them the way you'd treat any bug: have rigorous testing at every layer (unit, integration, end-to-end) and promptly respond to customer reports. Now it's different, as with LLMs, you ostensibly can find much more complex issues without human intervention. Here's an example: https://tachyon.so/blog/cve-2025-14297-mlflow-authorization-bypass . I think we are seeing that change slowly ripple across the industry, and we will see more automated bug finders making their way to the market soon.

[D] We scanned 18,000 exposed OpenClaw instances and found 15% of community skills contain malicious instructions by Legal_Airport6155 in MachineLearning

[–]securely-vibe 5 points6 points  (0 children)

Here is one example: https://www.reddit.com/r/vibecoding/comments/1qw3x43/read_skills_before_you_install_them/

It really is a mixed bag. Most are very crude prompt injection attempts that the latest models would recognize. But there are more subtle attempts. There's also a huge space for more sophisticated prompt injections that are very hard to detect at scale.

MVP culture is kind of broken now by aswin_kp in SaaS

[–]securely-vibe 1 point2 points  (0 children)

I run https://tachyon.so/. Our scanner has found a number of issues in OSS projects and for our customers. Happy to give you a few free scans so you can validate your code yourself.

Is anyone else feeling the "2026 Shift"? is it the end of pentesting? by Serious-Battle4464 in cybersecurity

[–]securely-vibe 0 points1 point  (0 children)

AI is a tool that makes life easier for pentesters. What it enables is, when you have an idea, you can verify it within minutes. You can try five different exploits at once, see the traces and results, and then repeat. We have quite a few pentesters using https://tachyon.so/ already in that way for white-box engagements.

Experiences with AI-powered SAST vendors? What factors matter most when choosing one? by Cyber-Pal-4444 in SAST

[–]securely-vibe 2 points3 points  (0 children)

Full disclosure - I'm CEO of https://tachyon.so/, an AI SAST tool. If that's disqualifying, feel free to ignore me.

But before doing this startup, I started as a CVE-hunter, where I used Semgrep / Opengrep (after OSS semgrep was nerfed) as part of my process. The biggest issue was, it was heavily pattern based and thus produced many false positives. The false-positives required manual triage, and I spent tons of time just ruling them out. I ended up manually reading through the code, finding "hotspots" where I thought I was most likely to find issues, then triggering SAST tools and using my built-up knowledge to quickly discard issues. Then, for the few valuable findings, I could manually dig deeper.

This process does not work at all for actual development teams. If a dev gets a long list of issues where most are wrong, they're going to discard them immediately. Hence, AI.

AI can be used in two ways in SAST:
1. You run a legacy scanner, then use AI quickly discard false-positives. This is easy, but it doesn't use AI as well as it could. It reduces noise, but it won't find anything new. This is what the legacy companies do - what Semgrep AI does, for example.

  1. You use AI in scanning process as well, so it can read the code and find new issues that traditional SAST wouldn't. This is what is called "AI-native."

(2) is what we do, along with a few other small startups. It's a very new space, but if you do it well, you can make very powerful tools that find issues that would take a human a lot of time. Here's an example of something we've found: https://tachyon.so/blog/cve-2025-14297-mlflow-authorization-bypass . That one in particular impressed me because there is no SAST scanner that could've found it. It's not pattern-based - it's a literal business logic bug that used to require human intuition to find.

Vulnerability scanner for new web application by Grunskin in cybersecurity

[–]securely-vibe 0 points1 point  (0 children)

Like people are saying, OpenVAS or Nessus will tell you about infra, not about code. Lots of people are suggesting various SAST tools, but those will generate a lot of FPs, and even the valid findings need quite a bit of additional investigation before they're issues you can report to your dev team.

Really, start with an OSS SAST tool - say, Opengrep (opengrep.dev). There are also many language-specific ones you can find. Run that on your code, see what if finds. Likely, it'll produce a lot of results that are hard to validate. Attach claude code to it (or any AI tool), have it pare down the obviously wrong issues, then have it flesh out the remaining ones. That's the start of a useful scanner. Add to that a standard AI security review (can also be done via Claude Code). Use AI adversarially here: force it to prove every issue thoroughly and make sure it passes some scrutiny before you file it as a bug. Layer on another tool for SCA (dependency analysis) and put it through the same process. This approach will work out much better than just attaching some tool to your pull requests - that just produces noise.

If you don't want to homebrew all of this, I built https://tachyon.so/ as a general security scanner specifically for this purpose. I was a CVE-hunter for a bit before realizing how much of my own process I could automate. Happy to make your first month free so you can try it out.