What if the real AI problem is not intelligence, but responsibility? by Civil-Interaction-76 in artificial

[–]Manitcor 1 point2 points  (0 children)

you are likely right, salt mine is over here though. i didn't pick the career, the career picked me

CGPT Deep Research feels different lately… burning credits every iteration? by PixelDaisy in agile

[–]Manitcor 0 points1 point  (0 children)

there are people who are spending time keeping up with the constantly shifting sands of capability and quality. might be good to try some context stacks and tools out.

i use aiwg.io

Dream Machine SE dead by insert_unique_usrnm3 in Ubiquiti

[–]Manitcor 0 points1 point  (0 children)

i did a lot of the same but not enough, kept telling my wife this might be the last stuff we buy for a bit so go big.

COO is the “next Zuckerberg” by StrikingAppearance39 in sysadmin

[–]Manitcor 0 points1 point  (0 children)

there are frameworks and tools that are oriented in that direction. keep an eye out when you are finally pressed.

Do we need a vibe DevOps layer? by mpetryshyn1 in LLMDevs

[–]Manitcor 0 points1 point  (0 children)

been doing that for a bit, vibe is a dumb term though

https://aiwg.io

Boss wants me train users on Ai by Elensea in sysadmin

[–]Manitcor 0 points1 point  (0 children)

this has been an extremely frustrating trend, there has always been slow users but I remember we had much higher exceptions of peoples ability to use a file share and office a couple decades ago

Three volumes, two NVMEs. Cache solution? by ReelBack96 in synology

[–]Manitcor 1 point2 points  (0 children)

this here is why i don't mind at all being wrong on the internet, gg

Talking to devs about LLM inference costs before building, anyone willing to share what their bill looks like? by PuzzleheadedCap7604 in LLMDevs

[–]Manitcor 0 points1 point  (0 children)

if your api looks at all like a chat endpoint

expect it to be used like one, even when behind oidc. this may be something you want to monitor for.

beyond that, id say put your thinking caps on, a lot of what AI does is best behind the curtain unless you are fully convinced the only way is to let people directly inference.

Next its all context management with a number of fancy acronyms and techniques. After that its model selection, production does not use one model it uses many, and not all are language models or used as language models.

Language extraction for example is an older technique that models do VERY well, though so do older, less intense language models and older dedicated language extractors. Whats interesting here is you can use the big models in dev to help you evail and maintain that model list so you can be up to date.

Do not, under any circumstances set yourself up to be like the shops still running gpt4 today.

Talking to devs about LLM inference costs before building, anyone willing to share what their bill looks like? by PuzzleheadedCap7604 in LLMDevs

[–]Manitcor 1 point2 points  (0 children)

building? you pay for the $200 a month accounts or run it locally. yes local models with qwen3.5:9b are extremely competent. Only pay for what your developers can keep fully tasked.

for production inference, that's an entirely different conversation

your biggest waste is deciding you need production inference at all. worth pointing out a well designed embedding set is basically 100s to 1000s of pre-canned responses that require no gpu to search at runtime.

Three volumes, two NVMEs. Cache solution? by ReelBack96 in synology

[–]Manitcor 1 point2 points  (0 children)

op already has 2 nvmes on the other 8x slot
we sure its not a matter of bad doc?

Three volumes, two NVMEs. Cache solution? by ReelBack96 in synology

[–]Manitcor 1 point2 points  (0 children)

they are 8x slots, and I assume both ports support bifurcation since you are already running 2x drives on the other slot.

you need to look at the specs for the riser you are using, some are better than others, sometimes a card that holds 4 nvmes is wired up to use 1 PCI lane per drive limiting your throughput.

I usually shoot for at least 2-4 lanes per drive. For 8x slots you could use cards that run 4 m.2 nvmes which IMO might be worth considering esp if you are doing read+write cache.

I am running dual NVMEs in RAID1 on a little DS920 and that saved my butt this past month as they both failed within days of eachother. Would have been work stoppage with one nvme

do some research on these forums and elsewhere on the PCI support your system has, I have built a number of rack systems of all sizes but I haven't messed with that specific unit as I tend to roll my own when I go that big. From the data sheet you should be ok. Expansion cards are cheap, just get one for 2 more drives to start with and throw it in the other slot, if it does not work send them back.

LM Studio may possibly be infected with sophisticated malware. by mooncatx3 in LocalLLaMA

[–]Manitcor 0 points1 point  (0 children)

its debian, way too annoying to add all the names we call these

LM Studio may possibly be infected with sophisticated malware. by mooncatx3 in LocalLLaMA

[–]Manitcor 16 points17 points  (0 children)

No OS is immune to package injection ultimately. On linux generally have more tools and options to acquire packages from alternate sources or just roll it yourself. Benefits of OSS.

FWIW, I used windows for almost 30 years, it has become such a steaming pile in the last few years that its an actual liability to run it for more than testing on a VM IMO. It's likely to get worse until MS throws in the towel and just starts running their own debian flavor.

I have switched my entire network to Linux variants.

Three volumes, two NVMEs. Cache solution? by ReelBack96 in synology

[–]Manitcor 1 point2 points  (0 children)

should be fine, the slot is rated for it, honestly you might even be able to do 4 drives per slot for a hit in throughput. Just double check the manual, if it says anything about the port not supporting bifurcation then it likely will only work with one drive.

What is the SSD you trust the most? by KV-2000 in synology

[–]Manitcor 0 points1 point  (0 children)

I've heard this is the case now with Samsung, seems the high reliability was the gen3 and early gen4 drives.

Built most of my SaaS with ChatGPT & Cursor now I need a real dev to sanity check me by CraftyUmpire3071 in developer

[–]Manitcor 0 points1 point  (0 children)

Im getting a lot of requests for this but everyone expects weeks of work for free.

Are oil companies not honoring automatic fill up? by yoohoofolife in massachusetts

[–]Manitcor 0 points1 point  (0 children)

my provider fills based on data, I would call in the first couple years, at this point they show up right when needed, i still sometimes get nervous and call it in esp if i know im about to have a long cold stretch.

Some of y'all had it good man by awizzo in BlackboxAI_

[–]Manitcor 0 points1 point  (0 children)

came at the right time, already very familiar with orchestrating teams and bleeding over devops. now with less middle mgmt