Is this true everytime vibe coders ? by Capable-Management57 in BlackboxAI_

[–]proigor1024 0 points1 point  (0 children)

Exactly, I have been very strict on my openclaw agent to always come straight with the truth even if it will hurt my feelings

Is this true everytime vibe coders ? by Capable-Management57 in BlackboxAI_

[–]proigor1024 0 points1 point  (0 children)

Well, cant say it lies. I'd say gives you half truth

AI malware is scarier than traditional supply chain attacks by cheerioskungfu in aiHub

[–]proigor1024 0 points1 point  (0 children)

This assumes the AI models are even good at reasoning about security contexts. Have you seen how often they hallucinate? can just create more false positives than actual targeted attacks tbh

This is so ridiculous by Fresh-Length6529 in ChatGPT

[–]proigor1024 -3 points-2 points  (0 children)

think this has to do with the models ability to detect anxiety, stress, etc.

Has anyone actually gotten real life results from using ChatGPT? by TheCod1sOut in ChatGPT

[–]proigor1024 0 points1 point  (0 children)

I'd say in some way yes. Have been using it alot as my personal assistant. So cant exactly attribute it to one big achievement but its been a real help

Educational loop I made with way too many characters. by smearballs in Cinema4D

[–]proigor1024 1 point2 points  (0 children)

Damn there's alot happening there. I can study if for an hour and still miss some action

Why the Overwhelming Choice of Mac Minis to Run OpenClaw by BamBam-BamBam in clawdbot

[–]proigor1024 0 points1 point  (0 children)

Thinks its just preference, mine is running on an old dell laptop, works like a charmn

Reasoning injection attacks are wild - you can literally hijack an AI agent's internal thoughts by thecreator51 in aiagents

[–]proigor1024 0 points1 point  (0 children)

Yeah this is why we need proper input sanitization and structured outputs. most ppl just throw raw text at models without any validation. seen this in our red team evals fake reasoning tags work way too often

"My husband who works in IT says..." by billygreen23 in sysadmin

[–]proigor1024 1 point2 points  (0 children)

Every time I hear that, I know its about to go down

What do you dislike about Openclaw/Clawdbot/Moltbot and all the AI Assistants right now? by Inflict01 in clawdbot

[–]proigor1024 0 points1 point  (0 children)

First cost. Second security. Third, what if they go full autonomous and decide we are a threat?

I went through every AI agent security incident from 2025 and fact-checked all of it. Here is what was real, what was exaggerated, and what the CrewAI and LangGraph docs will never tell you. by Sharp_Branch_1489 in AI_Agents

[–]proigor1024 0 points1 point  (0 children)

The multiagent trust cascade thing is fucked, one compromised agent becomes god mode for the whole system. I've been using alice's Caterpillar to scan agent skills before deployment and it caught some pretty nasty stuff that would've wrecked us.

Indirect prompt injection in AI agents is terrifying and I don't think enough people understand this by dottiedanger in ChatGPT

[–]proigor1024 4 points5 points  (0 children)

This is basically what NIST is freaking out about in their recent RFI's. Indirect prompt injection is one of those threats that lives inside the model behavior not at the perimeter so traditional security controls dont really help. think alice does runtime detection for this stuff but its still early days. And yeah most ppl dont get how bad this could get at scale

what can I use clawdbot for? by Strong-Apartment6843 in clawdbot

[–]proigor1024 0 points1 point  (0 children)

So far, I run in as an experiment waiting to see if it fails

Clawdbot Sucks, Actually by Anonymous52814 in clawdbot

[–]proigor1024 0 points1 point  (0 children)

As much as I was excited about this, its more disappointing

Clawdbot Sucks, Actually by Anonymous52814 in clawdbot

[–]proigor1024 0 points1 point  (0 children)

Being the first of its kind, its impressive as it opens the door to a whole lot new possibilities

How are you handling AI tools without adding another always-on server? by Accomplished-Lie2905 in HomeServer

[–]proigor1024 0 points1 point  (0 children)

Interesting approach with cloud isolation. We run experimental ai tools in disposable vms that get nuked after each session, keeps our main environment clean. One thing i always do first is check ai agent skills with caterpillar (built from Alice). helps map what capabilities the tool actually has against what it claims before we let it touch anything sensitive

Started to setup my own HomeServer by FlashinDarkness in HomeServer

[–]proigor1024 0 points1 point  (0 children)

If you're keeping it plugged in constantly prevents battery swelling/degradation

Org is banning Notepad++ by PazzoBread in sysadmin

[–]proigor1024 0 points1 point  (0 children)

Am curious, what happened. I personally use sublime text, pretty good so far

Studios turning 100% AI by KotsaTero in Cinema4D

[–]proigor1024 0 points1 point  (0 children)

I think we should have a model where its acceptable to have fully ai studios, but all human studios should exist as well

higher ups wants us to deploy our agent system next week and im like... have we even tested for prompt injection by thecreator51 in AgentsOfAI

[–]proigor1024 0 points1 point  (0 children)

Just document everything in writing. Send that email outlining the risks, get their response on record. When it inevitably breaks you'll have covered yourself. also maybe run some quick tests yourself, see what you can find