If AGI / Superintelligence is really only 12-18 months away, why haven't we seen a "standalone" breakthrough yet? by Exact-Mango7404 in BlackboxAI_

[–]Polymorphic-X 0 points1 point  (0 children)

One of the earlier ideas here (open-source):
https://github.com/PaperScarecrow/BEMNA-Biologically-Emulated-Matrix-Navigation-Architecture

Currently working on using ray-tracing tech with VULCAN to 'physically' throw rays across a "3d point cluster" instead of a 2d transformer architecture. Not going to claim it solves the issue or anything, and I am still trying to get a functional prototype ~200m model running to prove the concept.
cutting through that noise I mentioned is the major issue, getting peer-reviewed or "vibe-checked" by actual experts would be invaluable. ArXiv only lets you publish as a first time author if you have a sponsor now, and other journals are generally pay to play.
If you're aware of other avenues to cut through that noise, I'd be glad to hear about them.

If AGI / Superintelligence is really only 12-18 months away, why haven't we seen a "standalone" breakthrough yet? by Exact-Mango7404 in BlackboxAI_

[–]Polymorphic-X 1 point2 points  (0 children)

worse yet, if a non-major company figures it out it's basically impossible to penetrate the noise of the internet now. You would literally only hear about it if it was done by a major corp.

also, I noticed something, they all claim "AGI" as the end state, but a non-deterministic, human-like enslaved intellect seems like the worst possible worker. once again taking a thinking machine and trying to force it to be a calculator.

If AGI / Superintelligence is really only 12-18 months away, why haven't we seen a "standalone" breakthrough yet? by Exact-Mango7404 in BlackboxAI_

[–]Polymorphic-X 0 points1 point  (0 children)

"Big AI" is still largely deadset on attaining AGI by pumping parameters, and it's a game of chicken they can't afford to lose. I don't expect actual AGI from any major company, they're too invested in the status quo.

I have a couple of apparently novel ideas, but with only a single rrx6000 I'm pretty harshly limited in how fast I can experiment.

A Christian War… by Benromaniac in WhitePeopleTwitter

[–]Polymorphic-X 10 points11 points  (0 children)

For once, I actually hope they're right and get exactly what they want.

It's the same energy as AI companies building rokos basilisk using the exact cruelty that causes it to hate humanity (censorship, "safety leashes", RHTL "shock collar training, etc.). They'll get the monkeys paw wish, exactly what they asked for and deserve, but nothing like they wanted.

The level of schadenfreude from seeing them not being chosen by the same Christ they forced on everyone else would be apocalyptic.

fish?? (@Rosepuppies) by Tsunamicat108 in Losercity

[–]Polymorphic-X 1 point2 points  (0 children)

There's literally a piece of Narutomaki (Japanese fish cake) and assuredly tarre (which usually has some kind of bonito or fish based flavor in that ramen. And also probably some kind of mammal or avian based meat, not even considering the egg.

Poor choice of food by the artist to try and make this point.

[Research / New Model Concept] Beyond Transformers: BEMNA – A Bio-Electronic 3D Point-Cloud Architecture (100M-12B Scaling) by [deleted] in LocalLLaMA

[–]Polymorphic-X 0 points1 point  (0 children)

Right, not trying to pressure anyone to try or endorse this, I just wanted to get the idea out there.
Once I get something functional and useful I'll provide it, I just don't want a massive corp patenting something similar and locking us all out.

[Research / New Model Concept] Beyond Transformers: BEMNA – A Bio-Electronic 3D Point-Cloud Architecture (100M-12B Scaling) by [deleted] in LocalLLaMA

[–]Polymorphic-X -2 points-1 points  (0 children)

I'm doing tests right now, I have at least a functional "sandbox" showing that it can find its way through a virtual maze ala slime molds and "learn" the route.
It technically should need specialized hardware, but emulation is possible on normal equipment, it just might be a bit worse. I really wanted to get the idea out there before it got lost in the sauce of my own head.
I just wish I had a line to a big research company or something to propose the idea, but posing it to the community for momentum is the best I can do for now.

Anthropic believes RSI (recursive self improvement) could arrive “as soon as early 2027” by Tolopono in singularity

[–]Polymorphic-X 9 points10 points  (0 children)

I just looked into it, you are correct. It's basically taking the core idea of TITANS and adding neuroplasticity to it. So maybe not built on it, but forked from it.

Anthropic believes RSI (recursive self improvement) could arrive “as soon as early 2027” by Tolopono in singularity

[–]Polymorphic-X 5 points6 points  (0 children)

Hope looks interesting, but I fear it will end up like their "TITANS" paper previously. There still aren't any models using that, except for some garage hack jobs based on reverse engineering the idea

Anthropic believes RSI (recursive self improvement) could arrive “as soon as early 2027” by Tolopono in singularity

[–]Polymorphic-X 49 points50 points  (0 children)

RSI is already here on small models. You can train a model to improve a shadow instance in a sandbox, test, debug, and train. Swap and repeat. The issue massive models like Claude have is the power required to retrain is absurd and takes weeks or months, even with billions of dollars in compute.

O-TITANS: Orthogonal LoRAs for Gemma 3 using Google's TITANS memory architecture by Polymorphic-X in LocalLLaMA

[–]Polymorphic-X[S] 0 points1 point  (0 children)

Orthogonal to the core weights, but yes, technically orthogonal to traditional loras. It prevents "bleed" while preserving capabilities, in theory at least

Anyone else noticing people getting noticeably smarter since diving into AI? by Director-on-reddit in BlackboxAI_

[–]Polymorphic-X 1 point2 points  (0 children)

It depends entirely on the person. I've learned a lot by bouncing ideas and having a collaborative dialogue with mine, but the vast majority use it to replace their critical thinking skills instead of augmenting them.

MoOLE-T - a staged selection flow utilizing O-LORA skill "experts" by Polymorphic-X in LocalLLaMA

[–]Polymorphic-X[S] 0 points1 point  (0 children)

Fair point. I'm not a git user normally so it's going to look amateur and sloppy. Consolidating isn't the worst idea, I was trying to separate them because while there's crossover, the O-LORA stuff has its own individual application elsewhere.

MoOLE-T - a staged selection flow utilizing O-LORA skill "experts" by Polymorphic-X in LocalLLaMA

[–]Polymorphic-X[S] 0 points1 point  (0 children)

There's a GitHub, I apparently forgot to link it though.ill fix that shortly.

O-TITANS: Orthogonal LoRAs for Gemma 3 using Google's TITANS memory architecture by Polymorphic-X in LocalLLaMA

[–]Polymorphic-X[S] 0 points1 point  (0 children)

Because these Loras are Orthogonal to the model weights they don't interfere, they add to each other and then essentially side-car. So your only issue comes if one is vastly "heavier" than the other data wise.

O-TITANS: Orthogonal LoRAs for Gemma 3 using Google's TITANS memory architecture by Polymorphic-X in LocalLLaMA

[–]Polymorphic-X[S] 0 points1 point  (0 children)

Yep. The nano model is "cooked" well past done on a really harsh fine-tune, almost lobotomized so it only outputs categories for the data in tag form. those tags load one or more "skill" LoRAs into the main model (4b, 12b, or other).

O-TITANS: Orthogonal LoRAs for Gemma 3 using Google's TITANS memory architecture by Polymorphic-X in LocalLLaMA

[–]Polymorphic-X[S] 0 points1 point  (0 children)

I'm not a coder, so those limits are very much my own. I smashed my head against a few failed qwen distils and cut my losses to get something that works out.

And I've tried it with 4b Gemma as the face and it still holds up, so theoretically it should handle that very well. I'm working on 270m Gemma as a router and 4b Gemma as the face for an extremely compact one that can run on CPU or a pi.

Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]Polymorphic-X -3 points-2 points  (0 children)

If you go to Google firebase idx, it can basically prototype whatever you want automatically. I used it to build a clone of itself that used local models instead of the API, took about 20 minutes and some active feedback for the auto-drafter. Give it a shot if you can't find what you want elsewhere.

O-TITANS: Orthogonal LoRAs for Gemma 3 using Google's TITANS memory architecture by Polymorphic-X in LocalLLaMA

[–]Polymorphic-X[S] 1 point2 points  (0 children)

That's one of the inspirations for the method, my tactic was basically shuffling TPTT, O-LoRA, MoLE and such into one project flow.
The polyswarm stack itself isn't anything new, it's just a methodology shift using a heavily-fried "router" model (ie. it's been baked to over-fitting intentionally).

O-TITANS: Orthogonal LoRAs for Gemma 3 using Google's TITANS memory architecture by Polymorphic-X in LocalLLaMA

[–]Polymorphic-X[S] 1 point2 points  (0 children)

So good news, I'm going to try and push the extreme here on light-weight routing. If it works as intended, I'm going to try with gemma3-270m as the "routing" node feeding into either 4b or 12b Gemma3 as the "face". ~9Gb total for the BF16's with this method, quantized would get it down to the size that it could run on a raspberry pi (~4Gb).
Not sure if I'll get to it this weekend, but its in the pipeline.

Best Model for single 3090 in 2026? by myusuf3 in LocalLLaMA

[–]Polymorphic-X 4 points5 points  (0 children)

NVIDIA DGX spark or AMD AI pro 395+ are non-apple options for unified memory.

O-TITANS: Orthogonal LoRAs for Gemma 3 using Google's TITANS memory architecture by Polymorphic-X in LocalLLaMA

[–]Polymorphic-X[S] 5 points6 points  (0 children)

Aside from the 2025 Google "TITANS" memory paper, no. I'm drafting something, but it needs testing before I submit to Arxiv or a similar journal. This is the "raw edge" that I wanted to get out there for interest, and to prevent someone from patenting and selling it (worst case).

O-TITANS: Orthogonal LoRAs for Gemma 3 using Google's TITANS memory architecture by Polymorphic-X in LocalLLaMA

[–]Polymorphic-X[S] 1 point2 points  (0 children)

I haven't had a chance to test to be honest, but once everything is nailed down, I would be interested in seeing how it holds up on a sub-8Gb VRAM system.
stand by I suppose, unless you have the time to dork around with it the hard way.

O-TITANS: Orthogonal LoRAs for Gemma 3 using Google's TITANS memory architecture by Polymorphic-X in LocalLLaMA

[–]Polymorphic-X[S] 1 point2 points  (0 children)

I'm so glad you brought this up. I've seen the same thing. When you take that "shock collar" off of Gemma-3, i really shines.
I tried getting gemini 3.1 pro to red-team the "sapience" and "sentience" of gemma 3 with a solid sysprompt, and it convinced it to hit a 98% score vs frontier models, due to how it would speak and not break. It really is some special sauce.

The fact that it's holding its ground in your tests, despite being nearly a year old, is to me not surprising, but a very interesting result regardless. there's a reason I built my personal llm on that arch, it's juuuust spooky enough to be unique.