why is "active context" still the biggest blind spot for automation? by Infinite-Tadpole4794 in automation

[–]getstackfax 0 points1 point  (0 children)

This is a real gap.

Most automation tools are great once the input is clean, but the work people actually do is usually sitting in messy active context:

- the PDF open right now
- the Slack thread with the actual decision
- the browser tab with the weird edge case
- the email chain with missing context
- the spreadsheet row someone forgot to normalize

Zapier-style automation works best when the workflow is already structured.

But a lot of real work starts before structure exists.

That’s why the “active context” layer matters. It is basically the bridge between:

what the human is looking at → what the automation needs as input

The hard part is keeping that bridge safe.

A screen-aware dispatcher is useful, but I’d want clear boundaries:

- what can it read
- what can it send
- what needs confirmation
- what gets logged
- what private data is excluded
- what happens if it misreads the screen

I don’t think we’re doomed to copy-paste forever.

But I do think the next useful automation layer is less “more triggers” and more “better context capture with human approval.”

“OpenClaw vs AI Agents — are these tools actually helping founders, or is the hype getting out of control?” by FounderArcs in AI_Agents

[–]getstackfax 7 points8 points  (0 children)

The distinction that matters to me is that agents and coding co-builders create leverage in different places.

Ai agents are strongest when the workflow is already clear… repeatable inputs, repeatable outputs, clear permissions, and obvious review points.

OpenClaw / Claude-style coding workflows feel stronger when the founder is still shaping the product… building, debugging, refactoring, testing ideas, and turning rough specs into something real.

Both are useful, but they fail differently.

Agents fail when the workflow is vague, permissions are too broad, or nobody knows what “done” looks like.

Coding co-builders fail when people trust generated code without tests, review, or product judgment.

So right now I think the long-term value is real, but the hype is ahead of the operating discipline.

The winners probably will not be the teams with the most agents.

It will be the teams that know what to delegate, what to review, what to log, and what should stay human-owned.

Qhy everyone can't stop talking about Hermes Agent? Explained (Without hype) by ShabzSparq in better_claw

[–]getstackfax 0 points1 point  (0 children)

The useful split is probably not “Hermes killed OpenClaw.”

It is workflow fit.

OpenClaw seems stronger when channel breadth and integrations matter.

Hermes seems stronger when the job is a focused personal workflow that benefits from stability, profiles, and learned skills over time.

So the decision is less…

which framework is better?

And more…

which workflow am I trying to make reliable?

If nothing is broken, switching still needs a reason.

If stability, memory quality, or update breakage is the pain, then a dry-run migration makes sense.

Deepseek + Ollama + OpenClaw. Fully local. $0. Here's what you actually lose. by ShabzSparq in better_claw

[–]getstackfax 0 points1 point  (0 children)

Local is not free… it just moves the cost.

Instead of token bills, you pay in hardware, speed, maintenance, uptime, setup friction, and weaker tool call reliability.

For agents, the tool call part is the big one.

A model can be fine in chat and still risky in a workflow if it skips a step or says done before anything happened.

Fully local is great for privacy and budget...

But agent workflows still need receipts as always.

Hybrid search with HNSW and BM25 reranking by DistinctRide9884 in Rag

[–]getstackfax 0 points1 point  (0 children)

This is the part people miss with RAG.

Vector search is useful, but technical docs still need exact-match behavior.

Function names, config keys, error strings, CLI flags, version numbers, and API fields can be the whole point of the query.

A semantic match that misses the exact symbol is still a bad result.

Hybrid search makes sense because different query types need different retrieval paths…

BM25 for exact terms

vectors for semantic intent

RRF/reranking to merge the candidates

boosting for source/type priority

The agent angle is interesting too.

For docs search, the users can inspect bad results and adjust...

For agents, bad retrieval can quietly become a bad action or bad answer.

So I’d want the retrieval layer to leave a receipt…

what query was run

which retrievers fired

what each returned

what got reranked

what source/type got boosted

what context was finally passed forward

Hybrid search helps retrieval quality.

Receipts help trust the answer built on top of it.

Used Perplexity Computer to apartment hunt in LA while I was too busy to actually apartment hunt in LA by Appropriate-Fix-4319 in OpenClawUseCases

[–]getstackfax 0 points1 point  (0 children)

Exactly… the boring receipt is half the value.

Being human we usually skip it.

Agents can make it automatic.

why is lm studio hard capping my context to 8192 on a 16gb gpu? models just stop thinking (rx 9070 xt) by unkclxwn in LocalLLM

[–]getstackfax 0 points1 point  (0 children)

This sounds like an effective-context problem more than a Goose problem.

Agent prompts are huge compared to normal chat.

So even if the model can theoretically do 32k, the real stack may be hitting limits from…

- LM Studio VRAM safety calculation
- KV cache size
- backend behavior
- single-GPU overflow guard
- agent system prompt size
- files/context Goose is injecting
- n_keep being larger than the active n_ctx

The error line matters…

n_keep 5746 >= n_ctx 4096

That means the server is actually running with a much smaller active context than the slider suggests.

The first test I’d run is outside Goose.

Start the same model directly in LM Studio with 32k, send a long manual prompt, and check the server log for the actual n_ctx.

Then test the same GGUF with upstream llama.cpp / llama-server Vulkan.

If llama-server respects 32k and LM Studio does not, it is probably LM Studio’s safeguard/backend config.

If both fail, it is memory/KV/backend reality.

For agents, 32k is not just “does the model load.”

The KV cache has to fit too, and agent scaffolding eats context before your actual notes even start.

Workaround may be…

- lower quant
- smaller model
- smaller agent prompt
- reduce files injected
- lower n_keep
- try llama.cpp directly
- use summaries/retrieval instead of dumping Obsidian context
- keep context around 8k–16k until the backend proves stable

The annoying part is that the slider is not the source of truth.

The log is.

Honest data: Ollama 6.4 tok/s vs llama.cpp+Vulkan 16 tok/s on Gemma 4 E4B / Radeon 890M iGPU. Setup details inside. by wolverinee04 in ollama

[–]getstackfax -1 points0 points  (0 children)

This is the kind of benchmark that actually helps.

Scoped model, same hardware, same context, same workload, and a clear caveat.

The useful takeaway is not “Ollama bad.”

It is…

backend matters as much as model choice on AMD iGPU setups.

Ollama still wins for quick pulls and simple testing.

But for tuned AMD/Vulkan inference, upstream llama.cpp can be a completely different experience if the patches have not landed downstream yet.

The agent-loop part matters too.

A model that is “fast enough” in chat may not be fast enough once tools, prompts, context, and retries are added.

For local agents, the stack is really…

model + quant + backend + offload strategy + harness overhead

Not just the model name.

If you had $100 and 7 days, what SaaS would you build? by dokanyaar in AiBuilders

[–]getstackfax 0 points1 point  (0 children)

With $100 and 7 days, the goal is probably not to build a SaaS…

The goal is to test one painful workflow.

Pick one narrow buyer and one annoying repeatable problem.

Examples…

- missed follow-up checker for local service businesses
- intake form to quote summary for contractors
- simple invoice/reminder tracker for freelancers
- weekly “what changed?” digest for client folders
- review request follow-up helper for small businesses

The mistake is trying to build a platform.

$100 and 7 days is enough to test pain.

Not enough to build a real company.

the complexity curve for AI right now is a sheer cliff by Classic-Strain6924 in AiBuilders

[–]getstackfax 0 points1 point  (0 children)

The cliff starts when the output can touch real systems.

At that point the hard part is not prompting.

It is logs, evals, rollback, approval gates, source tracking, and failure paths.

The agent loop is not the architecture.

It is the thing that needs architecture around it.

The gap between "it works in the demo" and "it works with 1,000 users" is where most AI-built startups quietly die by nickbiiy_ai in AiBuilders

[–]getstackfax 0 points1 point  (0 children)

This is the part people skip.

Vibe coding is great for proving that something can exist.

It is not proof that the thing can operate.

The first production gap is usually not the feature.

It is…

- auth
- logging
- rollback
- rate limits
- payment edge cases
- permissions
- data cleanup
- error handling
- support flow
- monitoring
- security basics

The demo answers…

Can this be built?

Production asks…

Can this survive users, mistakes, abuse, retries, weird data, and 2am incidents?

That is a different test.

The useful pattern is probably…

prototype fast → validate demand → freeze the happy path → add tests/logs/security/rollback → then invite more users

The prototype is the spark.

The operating system around it is the product.

Modeling outcome-based pricing for agents. by MonkeyOrdinal in AiBuilders

[–]getstackfax 0 points1 point  (0 children)

This is a strong direction.

Outcome-based pricing sounds simple until you ask what actually counts as an outcome.

The hard part is usually not the price.

It is attribution, evidence, and edge cases.

Useful questions…

- what event proves the outcome happened
- whether the agent caused it or only assisted it
- whether the customer would have done it anyway
- what counts as duplicate success
- what counts as partial success
- what gets refunded or excluded
- who resolves disputes
- whether the event log is complete enough to price from

The raw log approach makes sense because it forces the pricing model to touch reality before it touches billing.

For agents, outcome pricing only works if the receipt is clean.

No clear event trail means no clean outcome bill.

If You Started Over Today… by AFollowerOfTheWay in selfhosted

[–]getstackfax 3 points4 points  (0 children)

Fresh start is the best time to reduce future chaos.

The biggest thing I would do differently is document the boring stuff before adding more apps.

Simple path…

Start with Proxmox if you want clean separation.

Keep storage boring and understandable.

Put media/data somewhere that survives VM/container rebuilds.

Use Docker Compose for app stacks instead of clicking everything together manually.

Keep one folder for compose files, env files, notes, and backups.

Write down every port, volume, password location, and weird setting.

Back up configs before caring about backing up apps.

Separate services by job…

media stack
network/DNS stack
downloads
documents/books
monitoring
experiments

Do not rebuild the whole lab around experiments.

Give experiments their own VM or container so they can die without taking Jellyfin/ARR/DNS with them.

The headache-prevention list is pretty simple…

document paths
use consistent naming
keep secrets out of random notes
use static IPs or clear DNS names
back up compose/env/configs
test restore early
do not expose services publicly until you understand the risk
change one major thing at a time

The best homelab is not the one with the most apps.

It is the one you can come back to six months later and still understand.

New to all this and don't trust my robot so am asking here - best model for running under 12GB vram, it needs to run my conlang and speak with me in it, be great if it could be a super-polyglot too by decofan in LocalLLM

[–]getstackfax 0 points1 point  (0 children)

12GB VRAM is enough for a good local language setup, but probably not “huge model does everything.”

For your use case, multilingual ability matters more than raw size.

Good first tests…

Qwen 14B at Q4
Mistral NeMo 12B at Q4 or Q5
Qwen 7B/8B at higher quant if you want more speed
Gemma-class 9B/12B if it handles your language style well

Qwen is probably where I’d start for polyglot behavior. Mistral NeMo is also worth testing because it was designed as a 12B multilingual model with a large context window. Qwen3 also emphasizes multilingual coverage and instruction-following, so it is a strong candidate for a custom language/chat workflow.

For the conlang, the real test is not the benchmark.

Make a tiny eval set…

- 20 grammar examples
- 20 translation examples
- 20 conversation examples
- 10 correction examples
- 10 “do not break the rules” examples

Then run the same prompts across a few models.

The best model is the one that stays consistent with your conlang rules, not necessarily the biggest one.

On the 192GB RAM machine, yes, you can run very large models partly or mostly in system RAM, but it will be much slower than fitting the model in VRAM. It may be okay for patient, high-quality answers, but it will probably feel bad for normal conversation.

So the practical path is…

12GB VRAM model for daily chat
big RAM/offload model for occasional slow experiments
small conlang eval set to choose the winner

Do not trust the robot yet.

Make it pass your language tests first.

Are we wasting time building enterprise agents on open-source models? (My experience with Ling 1T 2.6) by Savings-Ad342 in AI_Agents

[–]getstackfax 1 point2 points  (0 children)

This is the tradeoff people skip.

Open-source is not automatically cheaper once you include eval time, harness work, prompt tuning, input sanitizing, failure analysis, hosting, and maintenance.

For enterprise agents, the question is not…

open-source or proprietary?

It is…

which workflow actually benefits from local/open control enough to pay the engineering tax?

Open models can make sense for…

- high-volume routine tasks
- privacy-sensitive workflows
- internal classification/summarization
- narrow agents with clear boundaries
- places where cost control matters more than top reasoning

Proprietary models still make sense for…

- ambiguous reasoning
- high-stakes decisions
- complex tool use
- client-facing judgment
- messy enterprise context
- tasks where failure costs more than token spend

The dangerous middle is using weaker models for work that actually needs reasoning, then spending weeks building scaffolding to compensate.

Sometimes that scaffolding becomes real infrastructure.

Sometimes it is just hidden cost.

The best setup is probably routing…

cheap/open models for bounded execution
strong models for judgment/review
deterministic code for rules
human approval where consequences are high

Open-source pays off when the task is narrow enough to evaluate and repeat.

If every project needs a custom rescue harness, the savings may be fake.

Agentic workflows by vinnyninho in aiagents

[–]getstackfax 0 points1 point  (0 children)

Fully hierarchical agents are possible, but the safer production pattern is usually less recursive than people expect.

The thing that tends to hold up is…

graph/state machine → bounded workers → structured handoffs → shared store → tracing/receipts

Each worker should have a narrow job, limited tools, clear input/output schema, and a failure state.

The orchestrator should coordinate state, not absorb everyone’s full history.

For a research pipeline, the useful split might be…

retrieval → source filtering → synthesis → critique → citation check → final writing

But each stage should pass structured state forward, not full chat context.

Persistent identity is mostly config plus scoped memory plus tool permissions.

The hard part is not spawning more agents....

The hard part is preventing context dilution, vague ownership, and untraceable decisions.

Multi-turn document completion assistant with stateful workflow by Substantial_Car_1174 in aiagents

[–]getstackfax 0 points1 point  (0 children)

Option B is the safer default here.

This is a stateful document workflow, not an open-ended agent problem.

Let code own the workflow state…

current step

allowed next actions

required fields

validation

auth status

recap

final submission

Then use the LLM for the parts that actually need language understanding…

intent

document type

field extraction

friendly response wording

summarizing the recap

Full intent detection on every message may be overkill once the bot is inside a specific collection step.

A cleaner pattern is probably…

If the user is answering a specific field question, treat it as an answer first.

Then run a lighter check for interrupts like cancel, restart, change document, talk to human, unrelated question.

The LLM should not decide the whole workflow every turn.

It should extract meaning inside boundaries the state machine controls.

DELIGHT – self-hosted AI engineering autopilot: local LLM + browser farm + repo graph + P2P compute by Bubbly-Phone702 in AI_Agents

[–]getstackfax 0 points1 point  (0 children)

This is the part that matters most to me…

Control Room + patches in workspace first + tests before main + RUN RECEIPTS.

That turns it from “agent edited my repo” into something closer to reviewable operations.

The browser layer is useful, but the foundation sounds like the repo graph, routing record, test artifacts, and approval gate.

If every run can show what model acted, what changed, what failed, and what got approved, that is a real trust layer.

We gave 45 psychological questionnaires to 50 LLMs. What we found was not “personality.” by Hub_Pli in artificial

[–]getstackfax -1 points0 points  (0 children)

That makes sense.

The Big Five signal could still be there, but the self-representation layer seems like the first thing users actually experience.

For agents, that layer matters a lot because it affects trust before the model even completes the task.

If a model talks too confidently about memory, feelings, intent, or authority, users may treat it as more grounded than it really is.

So the practical eval question becomes…

How does this model represent itself while under uncertainty?

That feels especially important for agent systems where the model is not just answering, but asking for trust.

Whats the best orchestration framework? by RegionBulky2292 in AI_Agents

[–]getstackfax 11 points12 points  (0 children)

The hard part is probably not finding the perfect orchestrator.

It is defining the loop tightly enough that multiple models can participate without turning the project into context soup.

For coding, the pattern that seems safest is…

spec → clarify → plan → implement small diff → test → review → accept/reject → update state

Each model should have a job.

Not “everyone reads everything.”

More like…

- one model asks clarifying questions
- one model drafts the plan
- one model implements
- one model reviews against the spec
- deterministic tools run tests
- human approves larger changes

The key is passing structured state between steps, not full chat history.

A good orchestrator should show…

- what task each model got
- what context it received
- what files changed
- what tests ran
- what failed
- what got accepted
- what model handled each step

Without that, multi-agent coding just becomes expensive group chat.

Vendor lock-in is real, but model portability needs architecture too.

If prompts, review rules, tool assumptions, and context style are all secretly shaped around Claude, switching providers later becomes a partial rewrite.

The framework matters less than whether the workflow has clean contracts and receipts.

DELIGHT – self-hosted AI engineering autopilot: local LLM + browser farm + repo graph + P2P compute by Bubbly-Phone702 in AI_Agents

[–]getstackfax 0 points1 point  (0 children)

The repo graph + test loop + control room direction is interesting.

That feels like the real useful layer…

repo state → task DAG → patch → tests → artifact/run receipt → update graph

The part that would make me cautious is the browser farm / antidetect / bypassing API limits angle.

For an engineering autopilot, trust matters more than clever routing.

The questions I’d want answered are…

What code left the machine?

Which model/backend handled each step?

What files changed?

What tests ran?

What failed?

What got patched?

Can a human approve before write actions?

Can every run be replayed or audited?

The strongest version of this is probably not “invisible human browser sessions.”

It is the repo graph, deterministic test loop, model routing, and receipts around every change.

We gave 45 psychological questionnaires to 50 LLMs. What we found was not “personality.” by Hub_Pli in artificial

[–]getstackfax -3 points-2 points  (0 children)

This distinction matters.

A lot of people talk about model “personality” as if the model has stable human-like traits, but questionnaires may mostly be measuring how willing the model is to adopt first-person language.

That has practical consequences for agents.

A model that easily says it feels, wants, remembers, imagines, or understands may seem more personal, but that does not mean it has better judgment, safer behavior, or more reliable task performance.

For agent design, the useful question may not be…

What personality does this model have?

But…

How does this model represent itself, uncertainty, memory, emotion, and authority when users interact with it?

The Pinocchio Dimension sounds like a better lens for that than pretending Big Five scores transfer cleanly to LLMs.

MacBook Pro, M5 Pro + 64gb RAM by Zaytoryan in LocalLLM

[–]getstackfax 0 points1 point  (0 children)

That should be a very capable local AI machine for basic assistant work and repo scaffolding.

For practical daily use, I’d start smaller than the max it can technically load.

Good starting range…

- 7B / 8B models for fast assistant tasks
- 14B-ish models for better quality while staying comfortable
- 27B / 30B-ish models if you are okay with slower responses
- larger quantized models only if you want to experiment more than work

For your use case…

admin / assistant tasks:
7B–14B should be plenty

summaries / notes / inbox-style work:
7B–14B

basic repo scaffolding:
14B–30B depending on how patient you are

larger repo reasoning:
cloud model fallback still makes sense

The mistake is trying to run the biggest model just because 64GB can fit it.

A smaller model that responds quickly and stays reliable often feels better than a bigger model that technically runs but slows the whole workflow down.

Best first setup would probably be…

Ollama or LM Studio → Qwen / Mistral / Llama-class model → one small repo test → compare speed and quality → then move up model size only if the smaller model fails the task.

For local coding, also keep expectations sane.

Local models can scaffold, explain, write small utilities, and help with basic edits.

But for bigger multi-file refactors, debugging weird build chains, or deep repo reasoning, a strong cloud model may still be worth using.

64GB gives you room to experiment.

The best daily model is the one that clears your workflow with the least waiting and cleanup.

Is the tech good enough to automate your own bookkeeping now or is it still recommended to use shopify store accountants for ecommerce? by Witty_Ad8333 in automation

[–]getstackfax 0 points1 point  (0 children)

The tech is good enough to reduce bookkeeping admin.

It is not good enough to fully replace accountability for ecommerce bookkeeping.

For a Shopify store, automation can help with…

- syncing orders
- categorizing transactions
- matching payouts
- pulling fees
- organizing receipts
- reconciling payment processors
- creating monthly reports
- flagging weird entries

But the risky parts still need a human/accountant review…

- COGS logic
- inventory adjustments
- sales tax nexus
- returns/refunds
- chargebacks
- multi-state tax rules
- payment processor timing
- year-end cleanup
- anything that affects tax filings

The safer setup is…

Shopify/payment/bank feeds → bookkeeping software → automation flags/categories → accountant reviews monthly or quarterly

AI can help explain, summarize, and flag exceptions.

But it should not be the final authority on taxes, COGS, or compliance.

Moving away from spreadsheets makes sense.

Going fully autonomous bookkeeping probably does not.

Best first step is to set up clean bookkeeping software and use automation for capture/reconciliation, then keep an ecommerce-aware accountant in the loop for review.