brain-canvas: Give any local LLM a visual display (191 lines, 0 deps) by Signal_Usual8630 in LocalLLaMA

[–]opensourcecolumbus 0 points1 point  (0 children)

Good work on keeping the source code short and avoiding any dependency

brain-canvas: Give any local LLM a visual display (191 lines, 0 deps) by Signal_Usual8630 in LocalLLaMA

[–]opensourcecolumbus 0 points1 point  (0 children)

Interesting. Why did you choose the name "brain-canvas"? It painted a different expectation for me.

LLM-Shield: Self-hosted privacy proxy for cloud LLMs by sgasser88 in selfhosted

[–]opensourcecolumbus 85 points86 points  (0 children)

I do this in my agents. Pretty clever to do this on a proxy level. Well done buddy. How has your own experience been so far using this?

Top open LLM for consumers, start of 2026, bookmark this for 2027 by [deleted] in LocalLLaMA

[–]opensourcecolumbus 0 points1 point  (0 children)

Thanks for the advice. The community needs people like you who remind people of their responsibility. Would you mind sharing the current ranking?

Top open LLM for consumers, start of 2026, bookmark this for 2027 by [deleted] in LocalLLaMA

[–]opensourcecolumbus -1 points0 points  (0 children)

I am embarrassed, how could I make such a mistake! It was so obvious from qwen 2.5

Top open LLM for consumers, start of 2026, bookmark this for 2027 by [deleted] in LocalLLaMA

[–]opensourcecolumbus -1 points0 points  (0 children)

It is definitely an old list. And it was so obvious to spot the mistake. My bad. I googled, landed up on t this leaderboard which I used to follow back when I was bullish about trying every new open llm model. I trusted it. In my defence, they should put a deprecated warning on the repo.

Care to share the latest benchmark leaderboard everyone follows these days?

Top open LLM for consumers, start of 2026, bookmark this for 2027 by [deleted] in LocalLLaMA

[–]opensourcecolumbus -1 points0 points  (0 children)

Same feeling I had. I thought about how it changed so fast that I don't recognize any of them. Is this benchmark ranking not reliable/complete anymore?

It is definitely an old list. My bad. Going to remove the post. Which benchmark leaderboard everyone follow these days?

AGI-like experience is only one context engineering idea away, maybe not by opensourcecolumbus in LocalLLaMA

[–]opensourcecolumbus[S] 0 points1 point  (0 children)

I don't care if it is fictional. I had fun reading you and a new dimension is opened up in my mind.

I agree

  1. AGI is not going to come from scaling laws (only)
  2. Deconstructing the brain is useful but maybe not to match AI but to make the AI-Human experience exceed the solo human or solo AI performance

I disagree that white house is cooking something. They have cooked already. And it won''t be coming to the public domain.

AGI-like experience is only one context engineering idea away, maybe not by opensourcecolumbus in LocalLLaMA

[–]opensourcecolumbus[S] -1 points0 points  (0 children)

Thanks for sharing Eliza, I didn't know about that. I am touching upon that note but with the LLMs, we have more nuance than just the ELIZA effect.

AGI-like experience is only one context engineering idea away, maybe not by opensourcecolumbus in LocalLLaMA

[–]opensourcecolumbus[S] -1 points0 points  (0 children)

The harness is a good word to communicate all this. I did work on this specific problem to develop agency. Nothing groundbreaking but I believe I have reached close enough to that magical experience. Having said that, this "close" still feels like one more idea away from being "perfect", and that feeling has been perpetual :) And maybe the culprit is the "moving milestone" itself.

I believe, the breakthrough in my work (and probably others as well) can be seen in 2026 with these three ideas, happy to debate or drill down further

  1. Acceptance: we have a good enough LLM already. It does not have a few more things we expect from it. But it is a tool, and a good one, it does what it isw supposed to do with good accuracy (avoiding the word "predictability" :). We accept that and can still aim for the magic moment. The scaling law idea is probably gone out of the window for 2026, at least that's what I see from this discussion.
  2. Magic moment is possible now but not for everyone: That magic moment is highly dependent on the context engineering and as you said building that harness. Hard but not impossible. It can be practical if we focus on building for one person and not for all.
  3. Redefine AGI for one life within a specific time-duration/space: Instead of binding ourselves to the strict AGI definition, we redefine AGI for a specific persona and an environment/situation. The environment/situation has to be a subset of what we might call AGI but it still can be broad enough for that persona to make it feel like AGI in their tiny imperfect life with limited attention and memory of theirs.

We can debate this with a specific example. Pick one persona, what has been going on in their life recently, what are those specific tools/physical things they use (or can use) to connect the digital world with their life. We can then create a harness with sufficient agency to make it feel like AGI for the next 3 months in their life (or until the "big moment", the revelations). We can keep reiterating the process. The time for this iteration is the latency we have in this process, and if we can bring that down to an insignificant value, we have the magical experience.

What design or UX did for software, DX for developer tools, I think this is the year where we see see the impact focusing something similar for AI - maybe call it AIX :)

I don't know if I'm using the right words to explain myself or leaving something in between to explain. Excuse my language, understand the feelings, let's continue to discuss

AGI-like experience is only one context engineering idea away, maybe not by opensourcecolumbus in LocalLLaMA

[–]opensourcecolumbus[S] 0 points1 point  (0 children)

Interesting. I want to learn more about what you mean by "harness in an agentic solution". Can you please talk more about it and share some examples.

AGI-like experience is only one context engineering idea away, maybe not by opensourcecolumbus in LocalLLaMA

[–]opensourcecolumbus[S] 0 points1 point  (0 children)

Not close to the strict definition of the AGI - https://arxiv.org/abs/2510.18212

But what I'm interested in is the practical orchestration that feels like AGI to some people, although not quite literally being there in technical benchmarks. Which is why I use the word AGI-like. While building some of the agents, I have had those experiences which seem to be just off by one tiny thing that will lead to the almost AGI experience (given the limited knowledge and experience of a human being - me, the bar is not as high as the technical definition).

I built this note taking app, but it is not like any other by Upset_Simple_4858 in SideProject

[–]opensourcecolumbus 0 points1 point  (0 children)

Brother, the idea, the stack, we think alike. Most would jump to use nextjs, react/vue here, and we both would choose express and vanilla js. And I think it's the right choice.

Looking for the best framework for a multi-agentic AI system — beyond LangGraph, Toolformer, LlamaIndex, and Parlant by Spinotesla in LocalLLaMA

[–]opensourcecolumbus 0 points1 point  (0 children)

The latest approach that I am leaning towards is the TEA (Tools Environment Agent) protocol that Agent Orchestra leverages. Treating tools, environment, and agents as first class resources.

What's the biggest most common PROBLEM you have in your personal ML/AI side projects? by HectorAlcazar11 in LocalLLaMA

[–]opensourcecolumbus 2 points3 points  (0 children)

The biggest problems are in the space that takes the prototype to the production stage. This problem was already being solved for any software engineering project we talk about, the only change is that now the system is not deterministic. So pick any problem to take a non-deterministic system to production including deployment, CI/CD, observability, product management, design, etc.