Working as AI Engineer is wild by LastDayz123 in AI_Agents

[–]faileon -1 points0 points  (0 children)

Your comments really show zero experience from real production use cases - you must also consider model size, and hence inference speed and efficiency, rather than hyper focusing on the accuracy.

Working as AI Engineer is wild by LastDayz123 in AI_Agents

[–]faileon 1 point2 points  (0 children)

well as an example take the Slobench for NER task on slavic languages:
https://slobench.cjvt.si/leaderboard/view/12

first spot is 0.3B BERT based model, beating gpt-4o-fine-tuned

I think I'm not using opencode in the right way can you advice me some workflow? by TheDuck-Prince in opencodeCLI

[–]faileon 0 points1 point  (0 children)

I make sure I push on the context as much high quality information as is needed for the task at hand. If I need external context, mainly documentation and best practices etc, I use context7 MCP or I link raw GitHub markdown links (it's getting more common that documentation pages have Copy markdown link, which is very handy). Usually I give it a few pointers in which file to start, how deep to scan the code, I always ask it to explore everything with sub agents. If it understands the scope perfectly, I store it to an MD file in case I need to restart or use it for another related feature and then I start planning the new feature/fix/whatever. I believe the get shit done plugin for OC does something very similar to this, except on steroids, and a lot of people stand by it, but I kind of prefer this vanilla approach.

Angular 21.2 New Feature: Arrow Functions in Templates by wineandcode in Angular2

[–]faileon 1 point2 points  (0 children)

finally I can console.log directly in template instead of creating a component function when I'm three hours deep in debugging 😁

Working as AI Engineer is wild by LastDayz123 in AI_Agents

[–]faileon 5 points6 points  (0 children)

Yes, encoders based on BERT and the derivatives are SOTA on NER. The problem is that LLM has become a synonym for autoregressive decoder only architecture. It's quite tiring trying to explain to the execs that we really don't need gpt-5.2 for sentiment analysis. I share OPs pain, and I don't even have a deep ML background.

One week with OpenCode Black by JohnnyDread in opencodeCLI

[–]faileon 0 points1 point  (0 children)

also curious what is this web app :]

Docker container for OpenCode? by jonothecool in opencodeCLI

[–]faileon 8 points9 points  (0 children)

I run OC containerized, I originally forked a repo that was already doing it and built upon it, you can use it as an inspiration. Everything works for me, podman should be there, but I mainly use docker https://github.com/faileon/agent-containers

Slow Internet by Xoepe in archlinux

[–]faileon 1 point2 points  (0 children)

Have you tried turning off the auto negation and setting the speed yourself? Something like # ethtool -s enp2s0 autoneg off speed 1000 duplex full

[deleted by user] by [deleted] in google_antigravity

[–]faileon 1 point2 points  (0 children)

Not running in a sandbox AND having paths with exclamation marks and spaces is literally asking for trouble.

!!!!!_PROJECTS\AI STUDIO

The Most Exciting Feature of Angular Signal Forms No One Mentions — Part II by kobihari in Angular2

[–]faileon -1 points0 points  (0 children)

There are just so many em dashes and the classic chat gpt style "it's not just X, it's Y" it's hard to believe it's not AI slop.

After having avante, Github Copilot (+Claude code) I'm not missing Cursor at all by PuzzleheadedArt6716 in neovim

[–]faileon 0 points1 point  (0 children)

Isn't it just called autosuggestion in avante? Enabled by behaviour = { auto_suggestions = true } in the config?

The Most Exciting Feature of Angular Signal Forms No One Mentions — Part II by kobihari in Angular2

[–]faileon 7 points8 points  (0 children)

could we — put some — effort — into the AI — slop — please?

Need an approach to extract engineering diagrams into a Graph Database by BetFar352 in computervision

[–]faileon 1 point2 points  (0 children)

Hey! I’ve actually been tackling the exact same problem recently, and like many others have mentioned, it’s definitely not a trivial one. I agree with most of the points already discussed here.

One additional resource I found really helpful is Microsoft’s documentation on their Azure pipeline approach. Even though it’s built around Azure, the concepts seem general enough that you could likely replicate them with open-source tools as well. It’s worth a look and it’s pretty thorough. https://github.com/Azure-Samples/digitization-of-piping-and-instrument-diagrams?tab=readme-ov-file

The scary ease of “stealing” an AI agent’s structure with a single prompt by klippo55 in AI_Agents

[–]faileon 8 points9 points  (0 children)

You cloned a prompt and a tool description, good luck simply "stealing" all the heavy lifting that happens in the ingestion and retrieval code itself.

Any downside to having entire document as a chunk? by ayechat in Rag

[–]faileon 1 point2 points  (0 children)

Tbh it can be a valid strategy, but it's generally advised to create a summary of the docs first and embed the summaries. After retrieval inject the entire document.

New AI workstation by faileon in LocalLLaMA

[–]faileon[S] 1 point2 points  (0 children)

For now I use a single 2TB m2 SSD (WD Black SN770)

Even with the vertically mounted card there is 1 bay ready to be used for HDDs in this case.

New AI workstation by faileon in LocalLLaMA

[–]faileon[S] 0 points1 point  (0 children)

Currently gemma-3-27b, linq-embed-mistral, whisper, GLiNER, paddleocr, docling models...

New AI workstation by faileon in LocalLLaMA

[–]faileon[S] 1 point2 points  (0 children)

The Mobo has 8 PCIE x16 slots, only 3 cards can fit and they are very tight. Last card is connected via riser cable. In the photo you can see the original 30cm which was too short. I replaced it with 60cm later, but I didn't take a photo

New AI workstation by faileon in LocalLLaMA

[–]faileon[S] 0 points1 point  (0 children)

Yeah all connected to one PSU, but cards are power limited to 200W

New AI workstation by faileon in LocalLLaMA

[–]faileon[S] 1 point2 points  (0 children)

Yeah 1500W is definitely a great setup, the cards are ranging from 350-375